Skip to main content

Virtual Clusters: Isolated, Containerized HPC Environments in Kubernetes

  • Conference paper
  • First Online:
High Performance Computing. ISC High Performance 2022 International Workshops (ISC High Performance 2022)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13387))

Included in the following conference series:

Abstract

Today, Cloud and HPC workloads tend to use different approaches for managing resources. However, as more and more applications require a mixture of both high-performance and data processing computation, convergence of Cloud and HPC resource management is becoming a necessity. Cloud-oriented resource management strives to share physical resources across applications to improve infrastructure efficiency. On the other hand, the HPC community prefers to rely on job queueing mechanisms to coordinate among tasks, favoring dedicated use of physical resources by each application.

In this paper, we design a combined Slurm-Kubernetes system that is able to run unmodified HPC workloads under Kubernetes, alongside other, non-HPC applications. First, we containerize the whole HPC execution environment into a virtual cluster, giving each user a private HPC context, with common libraries and utilities built-in, like the Slurm job scheduler. Second, we design a custom Slurm-Kubernetes protocol that allows Slurm to dynamically request resources from Kubernetes. Essentially, in our system the Slurm controller delegates placement and scheduling decisions to Kubernetes, thus establishing a centralized resource management endpoint for all available resources. Third, our custom Kubernetes scheduler applies different placement policies depending on the workload type. We evaluate the performance of our system compared to a native Slurm-based HPC cluster and demonstrate its ability to allow the joint execution of applications with seemingly conflicting requirements on the same infrastructure with minimal interference.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 74.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. An open-source monitoring solution. https://prometheus.io/

  2. The apache software foundation. apache http server benchmarking tool. https://httpd.apache.org/docs/2.2/programs/ab.html

  3. VMware: The State of Kubernetes 2020. https://k8s.vmware.com/state-of-kubernetes-2020/

  4. Bailey, D., et al.: The nas parallel benchmarks. Int. J. High Perform. Comput. Appl. 5(3), 63–73 (1991)

    Google Scholar 

  5. Beltre, A.M., Saha, P., Govindaraju, M., Younge, A., Grant, R.E.: Enabling hpc workloads on cloud infrastructure using kubernetes container orchestration mechanisms. In: 2019 IEEE/ACM International Workshop on Containers and New Orchestration Paradigms for Isolated Environments in HPC (CANOPIE-HPC), pp. 11–20 (2019)

    Google Scholar 

  6. Cooper, B.F., Silberstein, A., Tam, E., Ramakrishnan, R., Sears, R.: Benchmarking cloud serving systems with YCSB. In: Proceedings of the 1st ACM Symposium on Cloud Computing, p. 143–154. SoCC 2010, ACM, New York, NY, USA (2010)

    Google Scholar 

  7. Delgado, P., Didona, D., Dinu, F., Zwaenepoel, W.: Job-aware scheduling in eagle: divide and stick to your probes. In: Proceedings of the Seventh ACM Symposium on Cloud Computing, pp. 497–509. SoCC 2016, ACM, New York, NY, USA (2016)

    Google Scholar 

  8. Felter, W., Ferreira, A., Rajamony, R., Rubio, J.: An updated performance comparison of virtual machines and linux containers. In: 2015 IEEE International Symposium on Performance Analysis of Systems and Software (ISPASS), pp. 171–172 (2015)

    Google Scholar 

  9. Herbein, S., et al.: Resource management for running hpc applications in container clouds, pp. 261–278, June 2016

    Google Scholar 

  10. Higgins, J., Holmes, V., Venters, C.: Orchestrating docker containers in the hpc environment, pp. 506–513, July 2015

    Google Scholar 

  11. Jin, T., Cai, Z., Li, B., Zheng, C., Jiang, G., Cheng, J.: Improving resource utilization by timely fine-grained scheduling. In: Proceedings of the Fifteenth European Conference on Computer Systems, pp. 1–16 (2020)

    Google Scholar 

  12. Li, M., Tan, J., Wang, Y., Zhang, L., Salapura, V.: Sparkbench: a comprehensive benchmarking suite for in memory data analytic platform spark. In: Proceedings of the 12th ACM International Conference on Computing Frontiers. CF 2015, ACM, New York, NY, USA (2015)

    Google Scholar 

  13. López-Huguet, S., Segrelles, J.D., Kasztelnik, M., Bubak, M., Blanquer, I.: Seamlessly managing HPC workloads through Kubernetes. In: Jagode, H., Anzt, H., Juckeland, G., Ltaief, H. (eds.) ISC High Performance 2020. LNCS, vol. 12321, pp. 310–320. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-59851-8_20

    Chapter  Google Scholar 

  14. Ortiz, J., Lee, B., Balazinska, M., Gehrke, J., Hellerstein, J.L.: Slaorchestrator: reducing the cost of performance slas for cloud data analytics. In: 2018 USENIX Annual Technical Conference (USENIX ATC 18), pp. 547–560. USENIX Association, Boston, MA, July 2018

    Google Scholar 

  15. Ousterhout, K., Canel, C., Ratnasamy, S., Shenker, S.: Monotasks: architecting for performance clarity in data analytics frameworks. In: Proceedings of the 26th Symposium on Operating Systems Principles, pp. 184–200 (2017)

    Google Scholar 

  16. Ousterhout, K., Wendell, P., Zaharia, M., Stoica, I.: Sparrow: distributed, low latency scheduling. In: Proceedings of the Twenty-Fourth ACM Symposium on Operating Systems Principles, pp. 69–84. ACM (2013)

    Google Scholar 

  17. Sfakianakis, Y., Marazakis, M., Bilas, A.: Skynet: performance-driven resource management for dynamic workloads. In: 2021 IEEE 14th International Conference on Cloud Computing (CLOUD). IEEE (2021)

    Google Scholar 

  18. Shvets, P., Voevodin, V., Nikitenko, D.: Approach to Workload Analysis of Large HPC Centers, pp. 16–30, July 2020

    Google Scholar 

  19. Zhao, L., et al.: Rhythm: component-distinguishable workload deployment in datacenters. In: Proceedings of the Fifteenth European Conference on Computer Systems, pp. 1–17 (2020)

    Google Scholar 

  20. Zhou, N., Georgiou, Y., Zhong, L., Zhou, H., Pospieszny, M.: Container orchestration on HPC systems. In: 2020 IEEE 13th International Conference on Cloud Computing (CLOUD), pp. 34–36 (2020)

    Google Scholar 

Download references

Acknowledgement

We thankfully acknowledge the support of the European Commission under the Horizon 2020 Programme through projects HiPEAC (GA-871174) and EVOLVE (GA-825061).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Antony Chazapis .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zervas, G., Chazapis, A., Sfakianakis, Y., Kozanitis, C., Bilas, A. (2022). Virtual Clusters: Isolated, Containerized HPC Environments in Kubernetes. In: Anzt, H., Bienz, A., Luszczek, P., Baboulin, M. (eds) High Performance Computing. ISC High Performance 2022 International Workshops. ISC High Performance 2022. Lecture Notes in Computer Science, vol 13387. Springer, Cham. https://doi.org/10.1007/978-3-031-23220-6_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-23220-6_24

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-23219-0

  • Online ISBN: 978-3-031-23220-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics