skip to main content
10.1145/1810085.1810116acmconferencesArticle/Chapter ViewAbstractPublication PagesicsConference Proceedingsconference-collections
research-article

InterferenceRemoval: removing interference of disk access for MPI programs through data replication

Published:02 June 2010Publication History

ABSTRACT

As the number of I/O-intensive MPI programs becomes increasingly large, many efforts have been made to improve I/O performance, on both software and architecture sides. On the software side, researchers can optimize processes' access patterns, either individually (e.g., by using large and sequential requests in each process), or collectively (e.g., by using collective I/O). On the architecture side, files are striped over multiple I/O nodes for a high aggregate I/O throughput. However, a key weakness, the access interference on each I/O node, remains unaddressed in these efforts. When requests from multiple processes are served simultaneously by multiple I/O nodes, one I/O node has to concurrently serve requests from different processes. Usually the I/O node stores its data on the hard disks, and different process accesses different regions of a data set. When there are a burst of requests from multiple processes, requests from different processes to a disk compete with each other for its single disk head to access data. The disk efficiency can be significantly reduced due to frequent disk head seeks.

In this paper, we propose a scheme, InterferenceRemoval, to eliminate I/O interference by taking advantage of optimized access patterns and potentially high throughput provided by multiple I/O nodes. It identifies segments of files that could be involved in the interfering accesses and replicates them to their respectively designated I/O nodes. When the interference is detected at an I/O node, some I/O requests can be re-directed to the replicas on other I/O nodes, so that each I/O node only serves requests from one or a limited number of processes. InterferenceRemoval has been implemented in the MPI library for high portability on top of the Lustre parallel file system. Our experiments with representative benchmarks, such as NPB BTIO and mpi-tile-io, show that it can significantly improve I/O performance of MPI programs. For example, the I/O throughput of mpi-tile-io can be increased by 105% as compared to that without using collective I/O, and by 23% as compared to that using collective I/O.

References

  1. M. Bhadkamkar, J. Guerra, L. Useche, S. Burnett, J. Liptak, R. Rangaswami, and V. Hristidis, "BORG: Block-reORGanization for Self-optimizing Storage Systems", In Proceedings of the 7th USENIX Conference on File and Storage Technologies, San Fancisco, CA, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. A. Ching, A. Choudhary, W. Liao, R. Ross, and W. Gropp, "Efficient Structured Data Access in Parallel File System", In Proceedings of IEEE International Conference on Cluster Computing, Hong Kong, China, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  3. A. Ching, A. Choudhary, K. Coloma, and W. Liao, "Noncontiguous I/O Accesses Through MPI-IO", In Proceedings of IEEE International Symposium on Cluster, Cloud, and Grid Computing, Tokyo, Japan, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Cluster File Systems, Inc. Lustre. "Lustre: A scalable, robust, highly-available cluster file system", http://www.lustre.org/. Online-document, 2010.Google ScholarGoogle Scholar
  5. H. Huang, W. Hung, and K. Shin, "FS2: Dynamic Data Replication in Free Disk Space for Improving Disk Performance and Energy Consumption", In Proceedings of ACM Symposium on Operating Systems Principles, Brighton, UK, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. W. Hsu, A. Smith, H. Young, "The Automatic Improvement of Locality in Storage Systems", ACM Transactions on Computer Systems, Volume 23, Issue 4, Nov. 2006, Pages 424--473. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. W. Hsu, A. Smith, H. Young, "The Automatic Improvement of Locality in Storage Systems", Technical Report CSD-03-1264, UC Berkeley, Jul. 2003.Google ScholarGoogle Scholar
  8. Interleaved or Random (IOR) benchmarks, http://www.cs.dartmouth.edu/pario/examples.html, Online-document, 2008.Google ScholarGoogle Scholar
  9. S. Iyer and P. Druschel, "Anticipatory scheduling: A disk scheduling framework to overcome deceptive idleness in synchronous I/O", In Proceedings of ACM Symposium on Operating Systems Principles, Banff, Canada, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. D. Kotz, "Disk-directed I/O for MIMD Multiprocessors.", ACM Transactions on Computer Systems, Volume 15, Issue 1, Feb. 1997, pages 41--74. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. S. Liang, S. Jiang, and X. Zhang, "STEP: Sequentiality and Thrashing Detection Based Prefetching to Improve Performance of Networked Storage Servers.", In Proceedings of International Conference on Distributed Computing Systems, Toronto, Canada, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Mpi-tile-io Benchmark, http: www-unix.mcs.anl.gov/thakur/pio-benchmarks.html. Online-document, 2009.Google ScholarGoogle Scholar
  13. M. Kandemir, S. Son, M. Karakoy, "Improving I/O Performance of Applications through Compiler-Directed Code Restructuring", In Proceedings of the 6th USENIX Conference on File and Storage Technologies, San Jose, CA, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. MPICH2, Argonne National Laboratory, http://www.mcs.anl.gov/research-/projects/mpich2/. Online-document, 2009.Google ScholarGoogle Scholar
  15. NAS Parallel Benchmarks, NASA AMES Research Center, http://www.nas.nasa.gov/Software/NPB/. Online-document, 2009.Google ScholarGoogle Scholar
  16. PVFS, http://www.pvfs.org. Online-document, 2010.Google ScholarGoogle Scholar
  17. P. Pacheco, "Parallel Programming with MPI", Morgran Kaufmann Publishers, pages 137--178, 1997. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. K. Seamons, Y. Chen, P. Jones, J. Jozwiak, and M. Winslett, "Server-directed collective I/O in Panda", In Proceedings of Supercomputing, San Diego, CA, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. F. Schmuck and R. Haskin, "GPFS: A shared-disk file system for large computing clusters.", In Proceedings of the 1st USENIX Conference on File and Storage Technologies, Monterey, CA, 2002, Monterey, CA, USA. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. R. Thakur, W. Gropp and E. Lusk, "Data Sieving and Collective I/O in ROMIO", In Proceedings of the 7th Symposium on the Frontiers of Massively Parallel Computation, Annapolis, MD, 1999. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. S3aSim I/O Benchmark, http://www-unix.mcs.anl.gov/thakur/s3asim.html. Online-document, 2009.Google ScholarGoogle Scholar
  22. The DiskSim Simulation Environment(v4.0), Parallel Data Lab, http://www.pdl.cmu.edu/DiskSim/. Online-document, 2009.Google ScholarGoogle Scholar
  23. Y. Wang and D. Kaeli, "Profile-Guided I/O Partitioning", In Proceedings of International Conference on Supercomputing, San Fancisco, CA, 2003. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. C. Wang, Z. Zhang, X. Ma, S. Vazhkudai, and F. Mueller, "Improving the Availability of Supercomputer Job Input Data Using Temporal Replication", In Proceedings of International Supercomputing Conference, Hamburg, Germany, 2009.Google ScholarGoogle ScholarCross RefCross Ref
  25. X. Zhang, S. Jiang, and K. Davis, "Making Resonance a Common Case: A High-performance Implementation of Collective I/O on Parallel File Systems", In Proceedings of IEEE International Parallel & Distributed Processing Symposium, Rome, Italy, 2009. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. InterferenceRemoval: removing interference of disk access for MPI programs through data replication

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    ICS '10: Proceedings of the 24th ACM International Conference on Supercomputing
    June 2010
    365 pages
    ISBN:9781450300186
    DOI:10.1145/1810085

    Copyright © 2010 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 2 June 2010

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article

    Acceptance Rates

    Overall Acceptance Rate584of2,055submissions,28%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader