Skip to main content

Toward Compiler Support for Scalable Parallelism Using Multipartitioning

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1915))

Abstract

Strategies for partitioning an application’s data play a fun- damental role in determining the range of possible parallelizations that can be performed and ultimately their potential efficiency. This paper describes extensions to the Rice dHPF compiler for High Performance Fortran which enable it to support data distributions based on multi- partitioning. Using these distributions can help close the substantial gap between the efficiency and scalability of compiler-parallelized codes for multi-directional line sweep computations and their hand-coded coun- terparts. We describe our the design and implementation of compiler support for multipartitioning and show preliminary results for a bench- mark compiled using these techniques.

This work has been supported by NASA Grant NAG 2-1181, sponsored by DARPA and Rome Laboratory, Air Force Materiel Command, USAF, under agreement number F30602-96-1-0159, and supported in part by the Los Alamos National Laboratory Computer Science Institute (LACSI) through LANL contract number 03891-99-23, as part of the prime contract (W-7405-ENG-36) between the Department of Energy and the Regents of the University of California. The U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright annotation thereon. The views and conclusions contained herein are those of the authors and should not be interpreted as representing the official policies or endorsements, either expressed or implied, of DARPA and Rome Laboratory or the U.S. Government.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. V. Adve, G. Jin, J. Mellor-Crummey, and Q. Yi. High Performance Fortran Compilation Techniques for Parallelizing Scientific Codes. In Proceedings of SC98: High Performance Computing and Networking, Orlando, FL, Nov 1998.

    Google Scholar 

  2. V. Adve and J. Mellor-Crummey. Using Integer Sets for Data-Parallel Program Analysis and Optimization. In Proceedings of the SIGPLAN ’98 Conference on Programming Language Design and Implementation, Montreal, Canada, June 1998.

    Google Scholar 

  3. S. Amarasinghe and M. Lam. Communication optimization and code generation for distributed memory machines. In Proceedings of the SIGPLAN ’93 Conference on Programming Language Design and Implementation, Albuquerque, NM, June 1993.

    Google Scholar 

  4. D. Bailey, J. Barton, T. Lasinski, and H. Simon. The NAS parallel benchmarks. International Journal of Supercomputing Applications, 5(3):63–73, Fall 1991.

    Article  Google Scholar 

  5. D. Bailey, T. Harris, W. Saphir, R. van derWijngaart, A. Woo, and M. Yarrow. The NAS parallel benchmarks 2.0. Technical Report NAS-95-020, NASA Ames Research Center, Dec.1995.

    Google Scholar 

  6. R. Bordawekar, A. Choudhary, K. Kennedy, C. Koelbel, and M. Paleczny. A model and compilation strategy for out-of-core data parallel programs. In Proceedings of the Fifth ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming, pages 1–10, Santa Barbara, CA, July 1995.

    Google Scholar 

  7. Z. Bozkus, L. Meadows, S. Nakamoto, V. Schuster, and M. Young. Compiling High Performance Fortran. In Proceedings of the Seventh SIAM Conference on Parallel Processing for Scientific Computing, pages 704–709, San Francisco, CA, Feb. 1995.

    Google Scholar 

  8. J. Bruno and P. Cappello. Implementing the beam and warming method on the hypercube. In Proceedings of 3rd Conference on Hypercube Concurrent Computers and Applications, pages 1073–1087, Pasadena, CA, Jan. 1988.

    Google Scholar 

  9. B. Chapman, P. Mehrotra, and H. Zima. Extending hpf for advanced data parallel applications. Technical ReportTR 94-7, Institute for Software Technology and Parallel Systems, University of Vienna, Austria, 1994.

    Google Scholar 

  10. N. Chrisochoides, I. Kodukula, and K. Pingali. Compiler and runtime support for irregular and adaptive applications. In Proceedings of the 1997 ACM International Conference on Supercomputing, pages 317–324, Vienna, Austria, July 1997.

    Google Scholar 

  11. M. Gerndt. Updating distributed variables in local computations. Concurrency:Practice and Experience, 2(3):171–193, Sept. 1990.

    Article  Google Scholar 

  12. S. K. S. Gupta, S. D. Kaushik, C. Huang, and P. Sadayappan. Compiling array expressions for efficient execution on distributed-memory machines. Journal of Parallel and Distributed Computing, 32(2):155–172, Feb. 1996.

    Article  Google Scholar 

  13. A. Müller and R. Rühl. Extending high performance fortran for the support of un-structure computations. Technical ReportTR-94-08, Ecole Polytechnique Fedérale de Zurich, November 1994.

    Google Scholar 

  14. N. Naik, V. Naik, and M. Nicoules. Parallelization of a class of implicit finite-difference schemes in computational uid dynamics. International Journal of High Speed Computing, 5(1):1–50, 1993.

    Article  Google Scholar 

  15. V. Naik. Performance effects of load imbalance in parallel CFD applications. In Proceedings of the Fifth SIAM Conference on Parallel Processing for Scientific Computing, 1992.

    Google Scholar 

  16. V. Naik. Scalability issues for a class of CFD applications. In Proceedings of the 1992 Scalable High Performance Computing Conference, Williamsburg, VA, Apr.1992.

    Google Scholar 

  17. V. Naik. A scalable implementation of the NAS parallel benchmark BT on distributed memory systems. IBM Systems Journal, 34(2), 1995.

    Google Scholar 

  18. N. Nedeljkovic and M. J. Quinn. Data-parallel programming on a network of heterogeneous workstations. Concurrency: Practice and Experience, 5(4):257–268,June 1993.

    Article  Google Scholar 

  19. R. F. Van derWijngaart. Efficient implementation of a 3-dimensional ADI method on the iPSC/860. In Proceedings of Supercomputing 1993, pages 102–111. IEEE Computer Society Press, 1993.

    Google Scholar 

  20. J. C. Yan, S. R. Sarukkai, and P. Mehra. Performance measurement, visualization and modeling of parallel and distributed programs using the aims toolkit. Software-Practice and Experience, 25(4):429–461, Apr. 1995.

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2000 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Chavarría-Miranda, D., Mellor-Crummey, J. (2000). Toward Compiler Support for Scalable Parallelism Using Multipartitioning. In: Dwarkadas, S. (eds) Languages, Compilers, and Run-Time Systems for Scalable Computers. LCR 2000. Lecture Notes in Computer Science, vol 1915. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-40889-4_21

Download citation

  • DOI: https://doi.org/10.1007/3-540-40889-4_21

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-41185-7

  • Online ISBN: 978-3-540-40889-5

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics