Skip to main content
Log in

Massive Storage Systems

  • Architechture
  • Published:
Journal of Computer Science and Technology Aims and scope Submit manuscript

Abstract

To accommodate the explosively increasing amount of data in many areas such as scientific computing and e-Business, physical storage devices and control components have been separated from traditional computing systems to become a scalable, intelligent storage subsystem that, when appropriately designed, should provide transparent storage interface, effective data allocation, flexible and efficient storage management, and other impressive features. The design goals and desirable features of such a storage subsystem include high performance, high scalability, high availability, high reliability and high security. Extensive research has been conducted in this field by researchers all over the world, yet many issues still remain open and challenging. This paper studies five different online massive storage systems and one offline storage system that we have developed with the research grant support from China. The storage pool with multiple network-attached RAIDs avoids expensive store-and-forward data copying between the server and storage system, improving data transfer rate by a factor of 2–3 over a traditional disk array. Two types of high performance distributed storage systems for local-area network storage are introduced in the paper. One of them is the Virtual Interface Storage Architecture (VISA) where VI as a communication protocol replaces the TCP/IP protocol in the system. VISA’s performance is shown to achieve better than that of IP SAN by designing and implementing the vSCSI (VI-attached SCSI) protocol to support SCSI commands in the VI network. The other is a fault-tolerant parallel virtual file system that is designed and implemented to provide high I/O performance and high reliability. A global distributed storage system for wide-area network storage is discussed in detail in the paper, where a Storage Service Provider is added to provide storage service and plays the role of user agent for the storage system. Object based Storage Systems not only store data but also adopt the attributes and methods of objects that encapsulate the data. The adaptive policy triggering mechanism (APTM), which borrows proven machine learning techniques to improve the scalability of object storage systems, is the embodiment of the idea about smart storage device and facilitates the self-management of massive storage systems. A typical offline massive storage system is used to backup data or store documents, for which the tape virtualization technology is discussed. Finally, a domain-based storage management framework for different types of storage systems is presented in the paper.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Gibson G A, Patterson D A. Designing disk arrays for high data reliability. Journal of Parallel and Distributed Computing, Jan, 1993, 17(1): 4–27.

    Article  Google Scholar 

  2. Gibson G. Scaling file service up and out. In Keynote Address at the 3rd USENIX Symp. File and Storage Technologies (FAST’04), San Francisco, California, March 31, 2004.

  3. Braam P J. The Lustre storage architecture. Technical Report, Cluster File Systems, Inc., January 2004, http://www.lustre.org/docs/lustre.pdf.

  4. Object Storage Architecture. White paper, January 2004, http://www.panasas.com/activescaleos.html.

  5. Intel Corporation. Object-based storage: The next wave of storage technology and devices. Jan. 2004, http://www.intel.com/labs/storage/osd/.

  6. Tierney B, Johnston W, Lee J et al. A data intensive distributed computing architecture for grid applications. Future Generation Computer Systems, April 2000, 16(5): 473–481.

    Article  Google Scholar 

  7. Baru C, Moore R, Rajasekar A et al. The SDSC storage resource broker. In Proc. CASCON’98 Conf., Toronto, Canada, 1998.

  8. Rhea S, Eaton P, Geels D et al. Pond: The OceanStore prototype. In Proc. the 2nd USENIX Conf. File and Storage Technologies (FAST’03), March 2003.

  9. Allcock W, Bester J, Bresnahan J et al. GridFTP protocol specification. GGF GridFTP Working Group Document, September 2002.

  10. Riedel E, Faloutsos C, Gibson G A, Nagle D. Active disks for large-scale data processing. IEEE Computer, June 2001, 34(6): 68–74.

    Google Scholar 

  11. Mesnier M, Thereska E, Ellard D et al. File classification in self-*storage systems. In Proc. the First Int. Conf. Autonomic Computing (ICAC-04), New York, NY. May 2004, pp. 44–51.

  12. Kephart J O, Chess D M. The vision of autonomic computing. IEEE Computer, 2003, pp. 41–50.

  13. Storage Networking Industry Association. CIM-SAN-1 Vendor Profile, Feb.2006, http://www.snia.org/tech_activities/SMI/CIMSAN/emc.doc

  14. Sivathanu M, Bairavasundaram L, Arpaci-Dusseau A C et al. Database-aware semantically-smart storage. In Proc. Fourth USENIX Symp. File and Storage Technologies (FAST’05), San Francisco, California, December 2005, pp. 239–252.

  15. Morris R. Storage: From atoms to people. In Keynote Address at Conference on File and Storage Technologies (FAST’02), Monterey, California, USA, January 28–30, 2002.

  16. Zhou K, Feng D, Wang F, Zhang J-L. Research on network RAID pipeline technology. Chinese Journal of Computers, Mar. 2005, 28(3): 319–325. (in Chinese)

  17. Farley M. Building Storage Network. 2nd Edition, Osborne/McGraw-Hill, May 22, 2001.

  18. Zeng L, Feng D, Zhou K, Wang F. Research on mechanism of network storage I/O pipelining. Mini-Micro Systems, 2006, 27(1): 42–45. (in Chinese)

  19. Wang F, Zhang J-L, Feng D et al. Adaptive control in Heter-RAID system. In Proc. Int. Conf. Machine Learning and Cybernetics, 4–5 Nov. 2002, pp. 842–845.

  20. Lampson B W. Hints for computer system design. ACM Operating Systems Review, October 1983, 15(5): 33–48.

    Google Scholar 

  21. KleinOsowski K, Ruwart T, Lilja D J. Communicating quality of service requirements to an object-based storage device. In Proc. 22nd IEEE/13th NASA Goddard Conf. Mass Storage Systems and Technologies (MSST 2005), Monterey, CA, USA, April 2005, pp. 224–231.

  22. Feng D, Zeng L. Adaptive policy triggering for load balancing. In Proc. the 6th Int. Conf. Algorithms and Architectures, Melbourne, Australia, October 2005, pp. 240–245.

  23. Feng D, Qin L, Zeng L et al. A scalable object-based intelligent storage device. In Proc. the 3rd Int. Conf. Machine Learning and Cybernetics, China, Aug. 2004, pp. 387–391.

  24. Wang F, Brandt S A, Miller E L, Long D D E. OBFS: A file system for object-based storage devices. In Proc. 21st IEEE/12th NASA Goddard Conf. Mass Storage Systems and Technologies (MSST’04), College Park, MD, April 2004.

  25. Morris J H, Satyanarayanan M, Conner M H et al. Andrew: A distributed personal computing environment. Communications of the ACM, Mar. 1986, 29(3): 184–201.

    Article  Google Scholar 

  26. Soltis S R, Ruwart T M, O’Keefe M T. The global file system. In Proc. the 5th NASA Goddard Conf. Mass Storage Systems and Technologies, College Park, MD, 1996.

  27. Satyanarayanan M, Kistler J J, Kumar P et al. Coda: A highly available file system for a distributed workstation environment. IEEE Trans. Computers, 1990, 39(4): 447–459.

    Article  Google Scholar 

  28. Schmuck F, Haskin R. GPFS: A shared-disk file system for large computing clusters. In Proc. the 2002 Conf. File and Storage Technologies (FAST’02), USENIX, Montery, CA, USA, Jan. 2002.

  29. Wang F, Zhang S, Feng D. A hybrid scheme for object allocation in a distributed object-storage system. In Proc. Int. Conf. Computational Science 2006 (ICCS’06), UK, 2006.

  30. Chen J, Feng D. VISA: A virtual interface storage architecture for improved network performance. In Proc. the 2nd Int. Conf. Embedded Software and Systems, Xi’an, China, December, 2005, pp. 587–592.

  31. Carns P H, Ligon W B III, Ross R B et al. PVFS: A parallel file system for Linux clusters. In Proc. the 4th Annual Linux Showcase and Conf., Atlanta, GA, Oct. 2000, pp. 317–327.

  32. Zhu Y, Jiang H et al. Design, implementation, and performance evaluation of a cost-effective fault-tolerant parallel virtual file system. In Proc. Int. Workshop on Storage Network Architecture and Parallel I/Os, New Orleans, LA, 2003.

  33. Barish B C, Weiss R. LIGO and the detection of gravitational waves. Physics Today, 1999, 52: 44.

    Google Scholar 

  34. Wulz C-E. CMS— Concept and physics potential. In Proc. II-SILAFAE, San Juan, Puerto Rico, 1998.

  35. NVO, 2004. http://www.us-vo.org/.

  36. Foster I, Kesselman C, Tuecke S. The anatomy of the grid: Enabling scalable virtual organizations. International Journal of High Performance Computing Applications, Aug. 2001, 15(3): 200–222.

    Article  Google Scholar 

  37. Chervenak A, Foster I, Kesselman C et al. The data grid: Towards an architecture for the distributed management and analysis of large scientific data sets. J. Network and Computer Applications, 2001, 23(3): 187–200.

    Article  Google Scholar 

  38. Rajasekar A, Jagatheesan A. Data grid management systems. In Proc. the 2003 ACM SIGMOD Int. Conf. Management of Data, April 2003.

  39. Jin H, Cortés T, Buyya R (eds.). High Performance Mass Storage and Parallel I/O. IEEE Press, John Wiley & Sons, Inc., 2002.

  40. Postel J, Reynolds J. File transfer protocol (FTP). Oct. 1985. http://www.ietf.org/rfc/rfc0959.txt?number=959.

  41. Postel J, Reynolds J. Telnet protocol specification (FTP). May 1983. http://www.ietf.org/rfc/rfc0854.txt?number=0854.

  42. Feng D, Zeng L, Wang F, Xia P. TLFS: High performance tape library file system for data backup and archive. In Proc. the 6th Int. Conf. High Performance Computing in Computational Sciences (VECPAR’2006), Rio de Janeiro, Brazil, July 10–12, 2006.

  43. Anderson T E, Dahlin M D, Neefe J M et al. Serverless network file systems. ACM Trans. Computer Systems, Feb. 1996, 14(1): 41–79.

    Article  Google Scholar 

  44. Lueth C. NetApp data protection: Double parity RAID for enhanced data protection with RAID-DP. March 5, 2005, http://www.netapp.com/tech_library/3298.html

  45. Palekar A, Ganapathy N, Chadda A et al. Design and implementation of a Linux SCSI target for storage area networks. In Proc. the 5th Annual Linux Showcase & Conference, Oakland, CA, USA, 2001.

  46. Mu F, Shu J, Li B et al. A virtual tape system based on storage area networks. In Proc. GCC’2004 Workshop on Storage Grid and Technologies, LNCS 3252, 2004, pp. 278–285.

  47. Myllymaki J, Livny M. Disk-tape joins: Synchronizing disk and tape access. ACM SIGMETRICS Performance Evaluation Review, 1995, pp. 279–290.

  48. ANSI. SCSI Stream Commands-2 (SSC-2). Revision 09, July 9, 2003, http://www.t10.org.

  49. Wheeler D. Any problem in computer science can be solved with another layer of indirection. March 26, 2005, http://www.computer.org/computer/homepage/0505/GEI/

  50. Storage Networking Industry Association (SNIA). (2003). SNIA Storage Management Initiative Specification Version 1.0.1. Retrieved October 5, 2005, from http://www.snia.org/smi/tech_activities/smi_spec_pr/spec/SMIS_v101.pdf.

  51. Distributed Management Task Force (DMTF). (Jun 14 1999). Common Information Model (CIM) Specification Version 2.2. Retrieved October 7, 2005, http://www.dmtf.org/standards/cim/DSP0004.pdf.

  52. Distributed Management Task Force (DMTF). Management Object Format (MOF). http://www.dmtf.org/education/mof.

  53. Distributed Management Task Force (DMTF). Specification for CIM operations over HTTP Version 1.0, DMTF Specification, Aug. 1999.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Dan Feng.

Additional information

Survey: This paper is supported by the National Natural Science Foundation of China under Grants No.60125208, No.60273074, No.60303032, No.69973017, and the National Grand Fundamental Research 973 Program of China under Grants No.2004CB318201, No. 2003CB317003.

Dan Feng received the Ph.D. degree from Huazhong University of Science and Technology (HUST), Wuhan, China, in 1997. She is currently a professor of School of Computer, HUST. Her research interests include computer architecture, storage system, parallel I/O, massive storage and performance evaluation.

Hai Jin received his Ph.D. degree in computer engineering from HUST in 1994. Dr. Jin now is a professor of computer science and engineering, the director of Cluster and Grid Computing Lab, and the Dean of School of Computer Science and Technology at HUST. He also is the chief scientist of the ChinaGrid project, one of the largest national grid computing in China. His research interests include computer architecture, cluster computing, grid computing, semantic web, peer-to-peer computing, network storage, network security, and pervasive computing.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Feng, D., Jin, H. Massive Storage Systems. J Comput Sci Technol 21, 648–664 (2006). https://doi.org/10.1007/s11390-006-0648-x

Download citation

  • Received:

  • Revised:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11390-006-0648-x

Keywords

Navigation