skip to main content
10.1145/2642769.2642782acmotherconferencesArticle/Chapter ViewAbstractPublication Pageseurompi-asiaConference Proceedingsconference-collections
research-article

Understanding the Memory-Utilization of MPI Libraries: Challenges and Designs in Implementing the MPI_T Interface

Published: 09 September 2014 Publication History

Abstract

The MPI Tools information interface (MPI_T), introduced as part of MPI 3.0 standard, has been gaining momentum in both the MPI and performance tools communities. In this paper, we investigate the challenges involved in profiling the memory utilization characteristics of MPI libraries that can be exposed to tools and libraries leveraging the MPI_T interface. We propose three design alternatives to enable such profiling from within MPI, and study their viability in light of these challenges. We analyze the benefits and shortcomings of each of them in detail, with a particular focus on the performance and memory overheads that they introduce. We evaluate the performance and scalability of these designs using micro-benchmarks, MPI-level benchmarks and applications. The overheads of the proposed design amounts to just 0.8% of the MILC application runtime with 4,096 processes. The paper also presents a case study that uses the MPI_T memory profiling information in MVAPICH2 to optimize the memory utilization of UH3D application runs, where memory savings of up to 7.3x was achieved.

References

[1]
OSU Micro-Benchmark Suite. http://mvapich.cse.ohio-state.edu/benchmarks/.
[2]
S. Gottlieb. MILC. In D. Padua, editor, Encyclopedia of Parallel Computing, pages 1130--1140. Springer US, 2011.
[3]
M. Jurenz, R. Brendel, A. Knüpfer, M. Müller, and W. E. Nagel. Memory allocation tracing with vampirtrace. In Proceedings of the 7th International Conference on Computational Science, Part II, ICCS '07, 2007.
[4]
M. Koop, T. Jones, and D. Panda. MVAPICH-Aptus: Scalable High-Performance Multi-Transport MPI over InfiniBand. In IPDPS, April 2008.
[5]
C. E. Leiserson, H. Prokop, and K. H. Randall. Using de Bruijn Sequences to Index a 1 in a Computer Word, 1998.
[6]
Mark Seager. Hyperion: A Titan of High-Performance Computing Systems, 2009.
[7]
MPI Forum. MPI 3.0 Standard. http://mpi-forum.org/docs/mpi-3.0/mpi30-report.pdf.
[8]
D. K. Panda, K. Tomko, K. Schulz, and A. Majumdar. The MVAPICH Project: Evolution and Sustainability of an Open Source Production Quality MPI Library for HPC. 2013.
[9]
SDSC. Gordon supercomputing system. http://www.sdsc.edu/us/resources/gordon/.
[10]
S. S. Shende and A. D. Malony. The tau parallel performance system. Int. J. High Perform. Comput. Appl., 20, 2006.
[11]
J. Vetter and C. Chambreau. mpip: Lightweight, scalable mpi profiling. http://mpip.sourceforge.net/.

Cited By

View all
  • (2018)Enabling callback-driven runtime introspection via MPI_TProceedings of the 25th European MPI Users' Group Meeting10.1145/3236367.3236370(1-10)Online publication date: 23-Sep-2018
  • (2016)Exploring the MPI tool information interfaceInternational Journal of High Performance Computing Applications10.1177/109434201560050730:2(212-222)Online publication date: 1-May-2016
  • (2015)Hardware-Centric Analysis of Network Performance for MPI ApplicationsProceedings of the 2015 IEEE 21st International Conference on Parallel and Distributed Systems (ICPADS)10.1109/ICPADS.2015.92(692-699)Online publication date: 14-Dec-2015

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
EuroMPI/ASIA '14: Proceedings of the 21st European MPI Users' Group Meeting
September 2014
183 pages
ISBN:9781450328753
DOI:10.1145/2642769
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

In-Cooperation

  • Kyoto University: Kyoto University
  • University of Tokyo
  • University of Tsukuba: University of Tsukuba

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 09 September 2014

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. MPI-3.0
  2. MPI_T
  3. MVAPICH
  4. memory-utilization
  5. profiling
  6. tools

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

EuroMPI/ASIA '14

Acceptance Rates

EuroMPI/ASIA '14 Paper Acceptance Rate 18 of 39 submissions, 46%;
Overall Acceptance Rate 18 of 39 submissions, 46%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)3
  • Downloads (Last 6 weeks)1
Reflects downloads up to 17 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2018)Enabling callback-driven runtime introspection via MPI_TProceedings of the 25th European MPI Users' Group Meeting10.1145/3236367.3236370(1-10)Online publication date: 23-Sep-2018
  • (2016)Exploring the MPI tool information interfaceInternational Journal of High Performance Computing Applications10.1177/109434201560050730:2(212-222)Online publication date: 1-May-2016
  • (2015)Hardware-Centric Analysis of Network Performance for MPI ApplicationsProceedings of the 2015 IEEE 21st International Conference on Parallel and Distributed Systems (ICPADS)10.1109/ICPADS.2015.92(692-699)Online publication date: 14-Dec-2015

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media