Skip to main content
Log in

Investigation of the parallel efficiency of a PC cluster for the simulation of a CFD problem

  • Original Article
  • Published:
Personal and Ubiquitous Computing Aims and scope Submit manuscript

Abstract

Previously, large-scale fluid dynamics problem required supercomputers, such as the Cray, and took a long time to obtain a solution. Clustering technology has changed the world of the supercomputer and fluid dynamics. Affordable cluster computers have replaced the huge and expansive supercomputers in computational fluid dynamics (CFD) field in recent years. Even supercomputers are designed in the form of clusters based on high-performance servers. This paper describes the configuration of the affordable PC hardware cluster as well as the parallel computing performance using commercial CFD code in the developed cluster. A multi-core cluster using the Linux operating system was developed with affordable PC hardware and low-cost high-speed gigabit network switches instead of Myrinet or Infiniband. The PC cluster consisted of 52 cores and easily expandable up to 96 cores in the current configuration. For operating software, the Rock cluster package was installed in the master node to minimize the need for maintenance. This cluster was designed to solve large fluid dynamics and heat transfer problems in parallel. Using a commercial CFD package, the performance of the cluster was evaluated by changing the number of CPU cores involved in the computation. A forced convection problem around a linear cascade was solved using the CFX program, and the heat transfer coefficient along the surface of the turbine cascade was simulated. The mesh of the model CFD problem has 1.5 million nodes, and the steady computation was performed for 2,000 time-integrations. The computation results were compared with previously published heat transfer experimental results to check the reliability of the computation. A comparison of the simulation and experimental results showed good agreement. The performance of the designed PC cluster increased with increasing number of cores up to 16 cores The computation (elapsed) 16-core was approximately three times faster than that with a 4-core.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Oracle Sun Grid Engine administration Guide Release 6.2 update 7

  2. Expo′sito RR, Taboada GL, Ramos S, Tourin˜o J, Doallo R (2012) Evaluation of messaging middleware for high-performance cloud computing. Pers Ubiquit Comput. doi:10.1007/s00779-012-0605-3

    Google Scholar 

  3. Chung K-Y, Yoo J, Kim KJ (2013) Recent trends on mobile computing and future networks. Pers Ubiquit Comput. doi:10.1007/s00779-013-0682-y

    Google Scholar 

  4. Kohtake N, Rekimoto J, Anzai Y (2001) Info Point: a device that provides a uniform user interface to allow appliances to work together over a network. Pers Ubiquit Comput 5:264–274

    Article  Google Scholar 

  5. Chang R-I, Chuang C-C (2012) A new spatial IP assignment method for IP-based wireless sensor networks. Pers Ubiquit Comput 16:913–928

    Article  Google Scholar 

  6. http://www.Beowulf.org

  7. Becker D, Savarese DF, Sterling T, Dorband JE, Ranawake UA, Packer CV (1995) Beowulf: a parallel workstation for scientific computation. In: Proceedings of the 24th International Conference on Parallel Processing, I. pp 11–14

  8. OSCAR: open source cluster application resources, http://www.openclustergroup.org

  9. Rocks cluster distribution, http://www.rocksclusters.org/Rocks/

  10. Han S, Goldstein R (2008) The heat/mass transfer analogy for a simulated turbine endwall with thermal boundary layer measurement and naphthalene sublimation method. Int J Heat Mass Transfer 51:3227–3244

    Article  MATH  Google Scholar 

  11. Han S, Goldstein R (2007) Thermal boundary layer measurement on turbine endwall and blade surface. Int J Heat Mass Transfer 129:1384–1394

    Google Scholar 

  12. Han S, Goldstein R (2008) The heat/mass transfer analogy for a simulated turbine blade with thermal boundary layer measurement and naphthalene sublimation method. Int J Heat Mass Transfer 51:5209–5225

    Article  MATH  Google Scholar 

  13. Kulkarni KS, Han S, Goldstein R (2011) Numerical simulation of thermal boundary layer profile measurement. Int J Heat Mass Transfer 47:869–877

    Article  Google Scholar 

  14. Sloan JD (2005) High performance Linux clusters with OACAR, Rocks, opemMosix and MPI O’Reilly

  15. Menter FR (1994) Two-equation eddy viscosity turbulence models for engineering applications. AIAA J 32:1598–1605

    Article  Google Scholar 

  16. Patankar SV (1980) Numerical heat transfer and fluid flow. McGRAW-HILL Book Company, New york

    MATH  Google Scholar 

  17. Kershaw DS (1978) The incomplete Cholesky-conjugate gradient method for the iterative solution of systems of linear equations. J Comput Phys 26:43–65

    Article  MATH  MathSciNet  Google Scholar 

  18. Snir M, Otto S, Huss-Lederman S, Walker D, Dongarra J (1996) MPI: The Complete Reference. The MIT Press, Cambridge

    Google Scholar 

  19. Carey GF, Shen Y, McLay RT (1998) Parallel conjugate gradient performance for least-squares finite elements and transport problems. Int J Numer Methods Fluids 28:1421–1440

    Article  MATH  Google Scholar 

  20. Choi HG, Kang SW, Yoo JY (2008) Parallel large eddy simulation of turbulent flow around MIRA model using linear equal-order finite element method—Parallel large eddy simulation around MIRA model. Int J Numer Meth Fluids 56:823–843

    Article  MATH  Google Scholar 

  21. http://www-users.cs.umm.edu/~karypis/metis

Download references

Acknowledgments

This work was supported by the IT R&D program of MSIP/KEIT. [10044910, Development of Multi-modality Imaging and 3D Simulation-Based Integrative Diagnosis-Treatment Support Software System for Cardiovascular Diseases].

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hyoung G. Choi.

Appendix

Appendix

1.1 The coefficients of the shear stress transport model

The set of empirical constants Φ = (σ k , σ ω , β, γ) used in the baseline model was calculated from two sets of constants Φ1 and Φ2 as follows:

$$\Upphi = \, F_{1} \Upphi_{1} + \left( {1 - F_{1} } \right)\Upphi_{2}$$
(4)

where the set of constants, Φ1, was derived from the original k − ω model such that

$$\sigma_{\text{k1}} = 0. 5,\;\sigma_{\omega 1} = 0. 5,\;\beta_{ 1} = 0.0 7 5,\;\beta^{*} = 0.0 9,\;\gamma_{1} = \beta_{1} /\beta^{*} - \frac{{\sigma_{\omega 1} \kappa^{2} }}{{\sqrt {\beta^{*} } }}$$
(5)

and the set of constants, k − ω, was derived from the standard k − ω model such that

$$\sigma_{\text{k2}} { = 1}.0,\;\sigma_{\omega 2} { = 1}. 8 5 6,\;\beta_{ 2} = 0.0 8 2 8,\;\beta^{*} = 0.0 9 ,\;\gamma_{2} = \beta_{2} /\beta^{*} - \frac{{\sigma_{\omega 2} \kappa^{2} }}{{\sqrt {\beta^{*} } }}.$$
(6)

F 1 can be expressed as

$$F_{1} = \tanh (\arg_{1}^{4} ),\;\arg_{1} = \hbox{min} \left[ {\hbox{max} \left( {\frac{\sqrt k }{0.00\omega y};\frac{500\nu }{{y^{2} \omega }}} \right);\frac{{4\rho \sigma_{\omega 2} k}}{{CD_{k\omega } y^{2} }}} \right]$$
(7)

where y is the distance to the next surface and CD is the positive portion of the cross-diffusion term

$$CD = \hbox{max} \left( {2\rho \sigma_{\omega } \frac{1}{\omega }\frac{\partial k}{{\partial x_{j} }}\frac{\partial \omega }{{\partial x_{j} }};10^{ - 20} } \right).$$
(8)

The eddy viscosity is defined as

$$\nu_{t} = \frac{{a_{1} k}}{{\hbox{max} (a_{1} \omega ;\Upomega F_{2} )}}$$
(9)

where Ω is the absolute value of the vorticity. F 2 can be expressed as

$$\begin{gathered} F_{2} = \tanh (\arg_{2}^{2} ) \hfill \\ \arg_{2} = \hbox{max} \left( {\frac{\sqrt k }{0.09\omega y};\frac{500\nu }{{y^{2} \omega }}} \right) \hfill \\ \end{gathered}.$$
(10)

Rights and permissions

Reprints and permissions

About this article

Cite this article

Han, S., Choi, H.G. Investigation of the parallel efficiency of a PC cluster for the simulation of a CFD problem. Pers Ubiquit Comput 18, 1303–1314 (2014). https://doi.org/10.1007/s00779-013-0733-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00779-013-0733-4

Keywords

Navigation