Skip to main content
Log in

A reinforcement learning framework for online data migration in hierarchical storage systems

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

Multi-tier storage systems are becoming more and more widespread in the industry. They have more tunable parameters and built-in policies than traditional storage systems, and an adequate configuration of these parameters and policies is crucial for achieving high performance. A very important performance indicator for such systems is the response time of the file I/O requests. The response time can be minimized if the most frequently accessed (“hot”) files are located in the fastest storage tiers. Unfortunately, it is impossible to know a priori which files are going to be hot, especially because the file access patterns change over time. This paper presents a policy-based framework for dynamically deciding which files need to be upgraded and which files need to be downgraded based on their recent access pattern and on the system’s current state. The paper also presents a reinforcement learning (RL) algorithm for automatically tuning the file migration policies in order to minimize the average request response time. A multi-tier storage system simulator was used to evaluate the migration policies tuned by RL, and such policies were shown to achieve a significant performance improvement over the best hand-crafted policies found for this domain.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Das R, Tesauro GJ, Walsh WE (2005) Model-based and model-free approaches to autonomic resource allocation. IBM Technical Report RC23802

  2. Hasinoff SW (2002) Reinforcement learning for problems with hidden state. Technical Report, University of Toronto, Department of Computer Science

  3. Howard RA (1960) Dynamic programming and Markov processes. Wiley, New York

    MATH  Google Scholar 

  4. Kretchmar RM, Anderson CW (1997) Comparison of CMACs and RBFs for local function approximators in reinforcement learning. In: Proceedings of the IEEE international conference on machine learning, Houston, TX, pp 834–837

  5. Lin LJ, Mitchell TM (1992) Memory approaches to reinforcement learning in non-Markovian domain. Carnegie Mellon School of Computer Science Technical Report CMU-CS-92-138

  6. Lu C, Alvarez GA, Wilkes J (2002) Aqueduct: online data migration with performance guarantees. In: Conference on file and storage technology (FAST’02), Monterey, CA. USENIX, Berkeley, pp 219–230

    Google Scholar 

  7. McCallum AK (1995) Reinforcement learning with selective perception and hidden state, PhD thesis, University of Rochester, Department of Computer Science

  8. Menon J, Pease DA, Rees B, Duyanovich LM, Hillsber BL (2003) IBM storage tank—a heterogeneous scalable SAN file system. IBM Syst J 42(2):250–267

    Article  Google Scholar 

  9. Meuleau N, Peshkin L, Kim K-E, Kaelbling LP (1999) Learning finite-state controllers for partially observable environments. In: Proceedings of the fifteenth conference on uncertainty in artificial intelligence (UAI), pp 427–436

  10. Paxson V, Floyd S (1995) Wide-area traffic: the failure of poisson modeling. IEEE/ACM Trans Netw 3(3):226–244

    Article  Google Scholar 

  11. Sinnwell M, Weikum G (1997) A cost-model-based online method for distributed caching. In: Proceedings of the thirteenth international conference on data engineering (ICDE), Birmingham, UK, pp 532–541

  12. Sutton RS, Barto AG (1998) Reinforcement learning: an introduction. MIT, Cambridge

    Google Scholar 

  13. Sun Microsystems Inc. (2004) Sun StorEdge QFS and SAM-FS software. Technical white paper. Available electronically at http://www.sun.com/storage/white-papers/qfs-samfs.pdf

  14. Tsitsiklis JN, Van Roy B (1997) An Analysis of Temporal-Difference Learning with Function Approximation. IEEE Trans Autom Control 42(5):674–690

    Article  MATH  Google Scholar 

  15. Vamplew P, Ollington R (2005) Global versus local constructive function approximation for on-line reinforcement learning. In: 18th Australian joint conference on artificial intelligence, Sydney, Australia, 5–9 December 2005. Lecture notes in computer science, vol 3809. Springer, Berlin, pp 113–121

    Google Scholar 

  16. Vengerov D (2005) Reinforcement learning framework for utility-based scheduling in resource-constrained systems. Sun Microsystems Laboratories Technical Report TR-2005-141

  17. Vengerov D (2005) A reinforcement learning approach to dynamic resource allocation. Sun Microsystems Laboratories Technical Report TR-2005-148

  18. Vengerov D (2006) Dynamic tuning of online data migration policies in hierarchical storage systems using reinforcement learning. Sun Microsystems Laboratories Technical Report TR-2006-157

  19. Verma A, Sharma U, Rubas J, Pease D, Kaplan M, Jain R, Devarakonda M, Beigi M (2005) An architecture for lifecycle management in very large file systems. In: Proceedings of the 22nd IEEE/13th NASA Goddard conference on mass storage systems and technologies (MSST), pp 160–168

  20. Wang L-X (1992) Fuzzy systems are universal approximators. In: Proceedings of the IEEE international conference on fuzzy systems (FUZZ-IEEE ’92), pp 1163–1169

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to David Vengerov.

Additional information

This material is based upon work supported by DARPA under Contract No. NBCH3039002.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Vengerov, D. A reinforcement learning framework for online data migration in hierarchical storage systems. J Supercomput 43, 1–19 (2008). https://doi.org/10.1007/s11227-007-0135-3

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-007-0135-3

Keywords