Skip to main content

The MuSE system: A flexible combination of on-stack execution and work-stealing

  • Conference paper
  • First Online:
Parallel and Distributed Processing (IPPS 1999)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 1586))

Included in the following conference series:

  • 116 Accesses

Abstract

Executing subordinate activities by pushing return addresses on the stack is the most efficient working mode for sequential programs. It is supported by all current processors, yet in most cases is inappropriate for parallel execution of indepented threads of control. This paper describes an approach of dynamically switching between efficient on-stack execution of sequential threads and off-stack spawning of parallel activities. The presented method allows to incorporate work-stealing into the scheduler, letting the system profit from its near-to-optimal loadbalancing properties.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. G. Acher, H. Hellwagner, W. Karl, and M. Leberecht. A PCI-SCI Bridge for Building a PC Cluster with Distributed Shared Memory. In Proceedings of the 6th International Workshop on SCI-Based High-Performance Low-Cost Computing, Santa Clara, CA, September 1996.

    Google Scholar 

  2. R. D. Blumofe and C. E. Leiserson. Scheduling Multithreaded Computations by Works Stealing. In Proceedings of the 35th Annual Sympsoium on Foundations of Computer Science (FOCS ’94), pages 356–368, Santa Fe, NM, USA, Nov 1994.

    Google Scholar 

  3. D. C. Cann. The Optimizing SISAL Compiler: Version 12.0. Technical Report UCRL-MA-110080, Lawrence Livermore, National Laboratory, April 1992.

    Google Scholar 

  4. P. Färber. Execution Architecture of the Multithreaded ADAM Prototype. PhD thesis, Eidgenössische Technische Hochschule, Zurich, Switzerland, 1996.

    Google Scholar 

  5. S. C. Goldstein, K. E. Schauser, and D. E. Culler. Lazy Threads: Implementing a Fast Parallel Call. Journal of Parallel and Distributed Computing, 37(1):5–20, 25 August 1996.

    Article  Google Scholar 

  6. H. Hellwagner, W. Karl, and M. Leberecht. Enabling a PC Cluster for High-Performance Computing. SPEEDUP Journal, June 1997.

    Google Scholar 

  7. M. Ibel, K. E. Schauser, C. J. Scheiman, and M. Weis High-Performance Cluster Computing Using Scalable Coherent Interface. In Proceedings of the 7th Workshop on Low-Cost/High-Performance Computing (SCIzzL-7), Santa Clara, USA, March 1997. SCIzzL.

    Google Scholar 

  8. A. M. Mainwaring and D. E. Culler Active Messages: Organization and Applications Programming Interface. Computer Science Division, University of California at Berkeley, 1995. http://now.cs.berkeley.edu/Papers/Papers/am-spec.ps.

    Google Scholar 

  9. J. Plevyak, V. Karamcheti, X. Zhang and A. Chien. A Hybrid Execution Model for Fine-Grained Languages on Distributed Memory Multicomputers. In Proodings of the 1995 ACM/IEEE Supercomputing Conference, San Diego, CA, December 1995. ACM/IEEE.

    Google Scholar 

  10. S. Skedzielewski and J. Glauert. IF1—An Intermediate Form for Applicative Languages. Technical Report TR M-170, Lawrence Livermore National Laboratory, July 1985.

    Google Scholar 

  11. IEEE Computer Society. IEEE Standard for Scalable Coherent Interface (SCI). The Institute of Electrical and Electronics Engineers, Inc., 345 East 47th Street, New York, NY 10017, USA, August 1993.

    Google Scholar 

  12. K. Taura, S. Matsuoka, and A. Yonezawa. Stack Threads: An Abstract Machine for Scheduling Fine-Grain Threads on Stock CPUs. In T. Ito and A. Yonezawa, editors, Proceedings of the International Workshop on the Theory and Practice of Parallel Programming, volume 907 of Lecture Notes of Computer Science, pages 121–136, Sendai, Japan, November 1994. Springer Verlag.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

José Rolim Frank Mueller Albert Y. Zomaya Fikret Ercal Stephan Olariu Binoy Ravindran Jan Gustafsson Hiroaki Takada Ron Olsson Laxmikant V. Kale Pete Beckman Matthew Haines Hossam ElGindy Denis Caromel Serge Chaumette Geoffrey Fox Yi Pan Keqin Li Tao Yang G. Chiola G. Conte L. V. Mancini Domenique Méry Beverly Sanders Devesh Bhatt Viktor Prasanna

Rights and permissions

Reprints and permissions

Copyright information

© 1999 Springer-Verlag

About this paper

Cite this paper

Leberecht, M. (1999). The MuSE system: A flexible combination of on-stack execution and work-stealing. In: Rolim, J., et al. Parallel and Distributed Processing. IPPS 1999. Lecture Notes in Computer Science, vol 1586. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0097889

Download citation

  • DOI: https://doi.org/10.1007/BFb0097889

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-65831-3

  • Online ISBN: 978-3-540-48932-0

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics