Abstract
Although the power of “yesterday’s” supercomputers is now available on the desktop, our yearning for even more computational power to solve even larger problems continues to grow. This paper covers the highlights of some of the author’s High Performance Computing (HPC) experiences dating back one of the first commercial supercomputers — Intel IPSC hypercube in the mid-80’s — through today where her current HPC work focuses on cluster computing using MPI.
The author will also give some of her prediction on where she thinks the HPC field is heading in the future. Her “Ant Theory” (even sugar ants display more complex planning and data processing than any human-built system of today) tells us we have a long way to go. Through newer technologies such as nanotubes and Bose-Einstein condensates, our future should even prove Moore’s law wrong, and let the HPC field tackle even bigger and more complex problems.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bell, G., Gray, J.: “What’s Next in High-Performance Computing?” in Communications of the ACM, 45, No. 2 (2002) 91–95.
Elster, A.C. and Reeves, A. P.: “Block-Matrix Operations Using Orthogonal Trees”, Proc. of the Third Conf. on Hypercube Systems and Applications, January 19–20, 1988 in Pasadena, CA, Ed. G. Fox, ACM, pp 1554–1561.
Elster, A.C. and Li, Hungwen: “Hypercube Algorithms on the Polymorphic Torus”, Proc. of the Fourth Conference on Hypercube, Concurrent Computers, and Applications, March 6–8, 1989 in Monterey, CA, Vol. I, Golden Gate Enterprises, pp 309–316. This paper was based on Cornell Computer Science TR 89-1003 and IBM Research Report, RJ 6775, 1989 (same title and authors).
Elster, A.C., Uyar, M.U., and Reeves, A.P.: “Fault Tolerant Matrix Operations on Hypercube Computers”, Proc. of the 1989 International Conf. on Parallel Processing, St. Charles, IL, August 8–12, 1989, Ed. F. Ris and P. M. Kogge, Penn State, Vol. III, pp 169–176.
Elster, A.C:, “Basic Matrix Subprograms for Distributed Memory Systems”, Proc. of the Fifth Distributed Memory Computing Conf. (DMCC5), in Charleston, SC, April 9–12, 1990, Ed. D. W. Walker and Q. Stout, IEEE Computer Society Press, pp 311–316. Received Student Paper Competition Award.
Elster, A.C. and Ramesh, P.S.: “Simulation of Charge Transport Using Parallel BLAS on the Intel Hypercube”, Xerox Internal Report X9200084, Webster Research Center, NY, April 1992.
Elster, A. C.: “Porting of the Parallel Pascal Translator from VAX 11/780 to IBM-PC/AT”. Master of Engineering Project Report, School of Electrical Engineering, Cornell University, Fall 1986.
Elster, A.C.: “Efficient Parallel Algorithms for Matrix Operations”, Master of Science Thesis, Cornell University, Aug. 1988.
Elster, A.C.: “Parallelization Issues and Particle-in-Cell Codes”, Ph.D. dissertation, Cornell University, August 1994. Abstract at: <http://www.englib.cornell.edu/thesesabstracts/August94/elster.html>
Birdsall, C.K. and Langdon, A.B.: Plasma Physics via Computer Simulations, Adam Hilger, Philadelphia, 1991.
Elster, A.C. and Presberg, David L.: “Setting Standards For Parallel Computing: The High Performance Fortran and Message Passing Interface Efforts”, Theory Center SMART NODE Newsletter, May, 1993, Vol.5, No. 3, Cornell University.
Hockney, R.W. and Jesshope, C.R.: Parallel Computers, Adam Hilger Ltd., Bristol, 1981.
Shaw, J.G.: Personal communications.
The Message Passing Interface Standard (MPI): <http://www-unix.mcs.anl.gov/mpi/index.html>
PETSc (Portable, Extensible Toolkit for Scientific Computation): <http://www-fp.mcs.anl.gov/petsc/>
Elster, A.C. and Liang, C.: “Developing and Testing Linear Solvers Using PETSc”, First SIAM Conference on Computational Science and Engineering, Washington, D.C., Sep 21–24, 2000. <http://www.siam.org/meetings/cse00/cp25.htm>
Sack, P. and Elster, A.C: “Fast MPI Broadcasts with Reliable Multicasting”, PARA’02, in this volume of Springer Verlag’s Lecture Notes on Computer Sciences, Fagerholm J. et al. (Eds), 2002.
“Top 500 Supercomputers”-Website maintained by Univ. of Mannheim, Germany and Univ. of Tennessee at Knoxville, U.S.A.. Yearly updates presented at the Supercomputing conferences. <http://www.top500.org>
Slides summarizing trend of Top 500 Supercomputers 1993-2001: <http://www.top500.org/slides/2001/11/>
Rotman, D: “The Nanotube Computer”, Technology Review, 105, No. 2, 36–45
“2001 Nobel Prizes”, Technology Review, 105, No. 2 (March 2002) 14–16
Technology Review, 105, No. 2. <http://www.technologyreview.com>
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Elster, A.C. (2002). High-Performance Computing: Past, Present, and Future. In: Fagerholm, J., Haataja, J., Järvinen, J., Lyly, M., Råback, P., Savolainen, V. (eds) Applied Parallel Computing. PARA 2002. Lecture Notes in Computer Science, vol 2367. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-48051-X_43
Download citation
DOI: https://doi.org/10.1007/3-540-48051-X_43
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-43786-4
Online ISBN: 978-3-540-48051-8
eBook Packages: Springer Book Archive