ABSTRACT
PNMPI extends the PMPI profiling interface to support multiple concurrent PMPI-based tools by enabling users to assemble tool stacks. We extend this basic concept to include new services for tool interoperability and to switch between tool stacks dynamically. This allows PNMPI to support modules that virtualize MPI execution environments within an MPI job or that restrict the application of existing, unmodified tools to a dynamic subset of MPI calls or even call sites.
Further, we extend PNMPI to platforms without dynamic linking, such as BlueGene/L, and we introduce an extended performance model along with experimental data from microbenchmarks to show that the performance overhead on any platform is negligible. More importantly, we provide significant new MPI tool components that are sufficient to compose interesting MPI tools. We present three detailed PNMPI usage scenarios that demonstrate that it significantly simplifies the creation of application-specific tools.
- Accelerated Strategic Computing Initiative. The ASCI sweep3d benchmark code. http://www.llnl.gov/asci_benchmarks/asci/limited/sweep3d/, December 1995.Google Scholar
- R. Bell, A. Malony, and S. Shende. ParaProf: A Portable, Extensible, and Scalable Tool for Parallel Performance Profile Analysis. In Proceedings of the International Conference on Parallel and Distributed Computing (Euro-Par 2003), pages 17--26, August 2003.Google ScholarCross Ref
- H. Brunst, D. Kranzlmüller, and W. Nagel. Tools for Scalable Parallel Program Analysis - Vampir NG and DeWiz. The International Series in Engineering and Computer Science, Distributed and Parallel Systems, 777:92--102, 2005.Google Scholar
- R. D. Falgout and U. M. Yang. hypre: a Library of High Performance Preconditioners. In Proceedings of the International Conference on Computational Science (ICCS), Part III, LNCS vol. 2331, pages 632--641, April 2002. Google ScholarDigital Library
- F. Gygi, E. W. Draeger, M. Schulz, B. R. de Supinski, J. A. Gunnels, V. Austel, J. C. Sexton, F. Franchetti, S. Kral, J. Lorenz, and C. W. Überhuber. Large-Scale Electronic Structure Calculations of High-Z Metals on the BlueGene/L Platform. In Proceedings of IEEE/ACM Supercomputing '06, November 2006. Google ScholarDigital Library
- B. Krammer, M. Müller, and M. Resch. Runtime Checking of MPI Applications with MARMOT. In Mini-Symposium on Tools Support for Parallel Programming at ParCo 2005, September 2005.Google Scholar
- G. Kumfert, J. Leek, and T. Epperly. Babel remote method invocation. In Proceedings of the 21st International Parallel and Distributed Processing Symposium, March 2007.Google ScholarCross Ref
- T. Ludwig, R. Wismüller, V. Sunderam, and A. Bode. OMIS --- On-line Monitoring Interface Specifi cation (Version 2.0), volume 9 of LRR-TUM Research Report Series. Shaker Verlag, Aachen, Germany, 1997. ISBN 3-8265-3035-7.Google Scholar
- J. May, D. Jefferson, N. Barton, R. Becker, J. Knap, G. Kumfert, J. Leek, and J. Tannahill. Introducing Cooperative Parallelism. Presented at the CCA Forum, presentation available at http://www.cca-forum.org/download/mtg/2007--01/may-coop-cca.ppt, January 2007.Google Scholar
- National Center for Atmospheric Research (NCAR). Community Climate System Model (CCSM). http://www.ccsm.ucar.edu/, 2006.Google Scholar
- M. Schulz and Bronis R. de Supinski. A Flexible and Dynamic Infrastructure for MPI Tool Interoperability. In Proceedings of the 2006 International Conference on Parallel Processing, August 2006. Google ScholarDigital Library
- J. S. Vetter and C. Chambreau. mpiP: Lightweight, Scalable MPI Profiling. http://www.llnl.gov/CASC/mpip/, April 2005.Google Scholar
- J. S. Vetter and B. R. de Supinski. Dynamic software testing of mpi applications with umpire. In Proceedings of IEEE/ACM Supercomputing '00, November 2000. Google ScholarDigital Library
- R. Vuduc, M. Schulz, D. Quinlan, B. de Supinski, and A. Sæbjørnsen. Improving Distributed Memory Applications Testing By Message Perturbation. In Proceedings of Parallel and Distributed Systems: Testing and Debugging (PADTAD), July 2006. Google ScholarDigital Library
- R. Wismüller. Interoperable Laufzeit-Werkzeuge für parallele und verteilte Systeme. Inaugural dissertation (Habilitation), Fakultät für Informatik, Technische Universität München, München, Germany, August 2001.Google Scholar
Recommendations
Tools-supported HPF and MPI parallelization of the NAS parallel benchmarks
FRONTIERS '96: Proceedings of the 6th Symposium on the Frontiers of Massively Parallel ComputationHigh Performance Fortran (HPF) compilers and communication libraries with the standardized Message Passing Interface (MPI) are becoming widely available, easing the development of portable parallel applications. The Annai tool environment supports ...
MPI + MPI: a new hybrid approach to parallel programming with MPI plus shared memory
Hybrid parallel programming with the message passing interface (MPI) for internode communication in conjunction with a shared-memory programming model to manage intranode parallelism has become a dominant approach to scalable parallel programming. While ...
MPI: past, present and future
PVM/MPI'07: Proceedings of the 14th European conference on Recent Advances in Parallel Virtual Machine and Message Passing InterfaceThis talk will trace the origins of MPI from the early message-passing, distributed memory, parallel computers in the 1980's, to today's parallel supercomputers. In these early days, parallel computing companies implemented proprietary message-passing ...
Comments