HeteroMPI: Towards a message-passing library for heterogeneous networks of computers

https://doi.org/10.1016/j.jpdc.2005.08.002Get rights and content

Abstract

The paper presents Heterogeneous MPI (HeteroMPI), an extension of MPI for programming high-performance computations on heterogeneous networks of computers. It allows the application programmer to describe the performance model of the implemented algorithm in a generic form. This model allows the specification of all the main features of the underlying parallel algorithm, which have an impact on its execution performance. These features include the total number of parallel processes, the total volume of computations to be performed by each process, the total volume of data to be transferred between each pair of the processes, and how exactly the processes interact during the execution of the algorithm. Given a description of the performance model, HeteroMPI tries to create a group of processes that executes the algorithm faster than any other group. The principal extensions to MPI are presented. We demonstrate the features of the library by performing experiments with parallel simulation of the interaction of electric and magnetic fields and parallel matrix multiplication.

Section snippets

Alexey Lastovetsky received a Ph.D. from the Moscow Aviation Institute in 1986, and a Doctor of Sciences degree from the Russian Academy of Sciences in 1997. His main research interests include algorithms, models and programming tools for high performance heterogeneous computing. He published over 60 technical papers in refereed journals and international conferences. He authored the monograph “Parallel computing on heterogeneous networks” published by Wiley in 2003. He is currently a lecturer

References (23)

  • M. Bhandarkar, L.V. Kale, E. de Sturler, J. Hoeflinger, object-based adaptive load balancing for MPI programs, in:...
  • Cited by (45)

    • Model-based selection of optimal MPI broadcast algorithms for multi-core clusters

      2022, Journal of Parallel and Distributed Computing
      Citation Excerpt :

      The target architecture for the presented method is a cluster of multi-core processors. While this is the most common architecture where MPI is used, there are other architectures such as heterogeneous clusters [23], [42] or Internet of Things [22], [15], where MPI either is used or can be used in the future. Such platforms are highly heterogeneous, represent a significant challenge for accurate modeling of MPI collective algorithms and are out of the scope of this paper.

    • Accelerating Communications in Federated Applications with Transparent Object Proxies

      2023, Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, SC 2023
    • Programming Reconfigurable Heterogeneous Computing Clusters Using MPI with Transpilation

      2020, Proceedings of H2RC 2020: 6th International Workshop on Heterogeneous High-Performance Reconfigurable Computing, Held in conjunction with SC 2020: The International Conference for High Performance Computing, Networking, Storage and Analysis
    View all citing articles on Scopus

    Alexey Lastovetsky received a Ph.D. from the Moscow Aviation Institute in 1986, and a Doctor of Sciences degree from the Russian Academy of Sciences in 1997. His main research interests include algorithms, models and programming tools for high performance heterogeneous computing. He published over 60 technical papers in refereed journals and international conferences. He authored the monograph “Parallel computing on heterogeneous networks” published by Wiley in 2003. He is currently a lecturer in the Department of Computer Science at University College Dublin, National University of Ireland. At UCD, he also leads Heterogeneous Computing Laboratory. He is on the editorial boards of the research journals “Parallel Computing” and “Programming and Computer Software”.

    Ravi Reddy is currently a Ph.D. student in the Department of Computer Science at University College Dublin, National University of Ireland. His main research interests are design of algorithms and tools for parallel and distributed computing systems.

    View full text