Skip to main content
Log in

The design and implementation of MPI collective operations for clusters in long-and-fast networks

  • Published:
Cluster Computing Aims and scope Submit manuscript

Abstract

Several MPI systems for Grid environment, in which clusters are connected by wide-area networks, have been proposed. However, the algorithms of collective communication in such MPI systems assume relatively low bandwidth wide-area networks, and they are not designed for the fast wide-area networks that are becoming available. On the other hand, for cluster MPI systems, a bcast algorithm by van de Geijn, et al. and an allreduce algorithm by Rabenseifner have been proposed, which are efficient in a high bi-section bandwidth environment. We modify those algorithms so as to effectively utilize fast wide-area inter-cluster networks and to control the number of nodes which can transfer data simultaneously through wide-area networks to avoid congestion. We confirmed the effectiveness of the modified algorithms by experiments using a 10 Gbps emulated WAN environment. The environment consists of two clusters, where each cluster consists of nodes with 1 Gbps Ethernet links and a switch with a 10 Gbps upper link. The two clusters are connected through a 10 Gbps WAN emulator which can insert latency. In a 10 millisecond latency environment, when the message size is 32 MB, the proposed bcast and allreduce are 1.6 and 3.2 times faster, respectively, than the algorithms used in existing MPI systems for Grid environment.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Allman, M., Kruse, H., Ostermann, S.: An application-level solution to TCP’s satellite inefficiencies. In: Proc. of the First Intl. Workshop on Satellite-based Information Services (WOSBIS) (1997)

  2. Barnett, M., Gupta, S., Payne, D.G., Shuler, L., van de Geijn, R., Watts, J.: Interprocessor collective communication library (InterCom). In: Proc. of the Scalable High Performance Computing Conference, pp. 357–364 (1994)

  3. Barnett, M., Gupta, S., Payne, D.G., Shuler, L., van de Geijn, R., Watts, J.: Building a high-performance collective communication library. In: Proc. of the 1994 Conference on Supercomputing (SC94), pp. 107–116 (1994)

  4. den Burger, M., Kielmann, T., Bal, H.E.: Balanced multicasting: high-throughput communication for grid applications. In: Proc. of the 2005 ACM/IEEE Conference on Supercomputing (SC’05) (2005)

  5. Chan, E., van de Geijn, R., Gropp, W., Thakur, R.: Collective communication on architectures that support simultaneous communication over multiple links. In: Proc. of the ACM SIGPLAN Symp. on Principles and Practice of Parallel Programming (PPoPP 2006), pp. 2–11 (2006)

  6. Gabriel, E., Resch, M., Rühle, R.: Implementing MPI with optimized algorithms for metacomputing. In: Proc. of the Third MPI Developer’s and User’s Conference (MPIDC’99), pp. 31–41 (1999)

  7. GLIF: Global Lambda Integrated Facility, http://www.glif.is

  8. GridMPI Project, http://www.gridmpi.org

  9. GtrcNET-10, http://www.gtrc.aist.go.jp/gnet

  10. Ishikawa, Y.: YAMPII official home page, http://www.il.is.s.u-tokyo.ac.jp/yampii

  11. Karonis, N.T., de Supinski, B.R., Foster, I.T., Gropp, W., Lusk, E.L., Lacour, S.: A multilevel approach to topology-aware collective operations in computational grids. Tech. Rep. ANL/MCS-P948–0402 (2002)

  12. Kielmann, T., Bal, H.E., Gorlatch, S.: Bandwidth-efficient collective communication for clustered wide area systems. In: Proc. of the 14th Intl. Parallel and Distributed Processing Symp., pp. 492–499 (1999)

  13. Kodama, Y., Kudoh, T., Takano, R., Sato, H., Tatebe, O., Sekiguchi, S.: GNET-1: gigabit Ethernet network testbed. In: IEEE Intl. Conf. on Cluster Computing (Cluster2004), pp. 185–192 (2004)

  14. Matsuda, M., Ishikawa, Y., Kudoh, T.: Evaluation of MPI implementations on grid-connected clusters using an emulated WAN environment. In: 3rd Intl. Symp. on Cluster Computing and the Grid (CCGrid2003), pp. 10–17 (2003)

  15. Matsuda, M., Ishikawa, Y., Kudoh, T., Kodama, Y., Takano, R.: TCP adaptation for MPI on long-and-fat networks. In: IEEE Intl. Conf. on Cluster Computing (Cluster2005), pp. 1–10 (2005)

  16. Rabenseifner, R.: Automatic MPI counter profiling of all users: first results on a CRAY T3E 900-512. In: Proc. of the Message Passing Interface Developers and Users Conference 1999 (MPIDC99), pp. 77–85 (1999)

  17. Rabenseifner, R.: Optimization of collective reduction operations. In: Intl. Conf. on Computational Science, LNCS 3036, pp. 1–9. Springer (2004)

  18. Takano, R., Kudoh, T., Kodama, Y., Matsuda, M., Tezuka, H., Ishikawa, Y.: Design and evaluation of precise software pacing mechanisms for fast long-distance networks. In: 3rd Intl. Workshop on Protocols for Fast Long-Distance Networks (PFLDnet05) (2005)

  19. TeraGrid, http://www.teragrid.orgVol

  20. Thakur, R., Rabenseifner, R., Gropp, W.: Optimization of collective communication operations in MPICH. Int. J. High Perform. Comput. Appl. 19(1), 49–66 (2005)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Motohiko Matsuda.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Matsuda, M., Kudoh, T., Kodama, Y. et al. The design and implementation of MPI collective operations for clusters in long-and-fast networks. Cluster Comput 11, 45–55 (2008). https://doi.org/10.1007/s10586-007-0050-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10586-007-0050-7

Keywords

Navigation