Skip to main content

Advertisement

Log in

Boosted regression for predicting CPU utilization in the cloud with periodicity

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

Predicting CPU usage is crucial to cloud resource management. Precise CPU prediction, however, is a tough challenge due to the variable and dynamic nature of CPUs. In this paper, we introduce TrAdaBoost.WLP, a novel regression transfer boosting method that employs Long Short-Term Memory (LSTM) networks for CPU consumption prediction. Concretely, a dedicated Periodicity-aware LSTM (PA-LSTM) model is specifically developed to take into account the use of periodically repeated patterns in time series data while making predictions. To adjust for variations in CPU demands, multiple PA-LSTMs are trained and concatenated in TrAdaBoost.WLP using a boosting mechanism. TrAdaBoost.WLP and benchmarks have been thoroughly evaluated on two datasets: 160 Microsoft Azure VMs and 8 Google cluster traces. The experimental results show that TrAdaBoost.WLP can produce promising performance, improving by 32.4% and 59.3% in terms of mean squared error compared to the standard Probabilistic LSTM and ARIMA.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Algorithm 1
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Data Availability Statement

This research uses the Azure Public Dataset V1 [14] and Google cluster traces [15]. The processed data is publicly available on this GitHub repository [43].

References

  1. Dogani J, Khunjush F, Seydali M (2023) Host load prediction in cloud computing with discrete wavelet transformation (dwt) and bidirectional gated recurrent unit (bigru) network. Comput Commun 198:157–174

    Article  Google Scholar 

  2. Wu Y, Hwang K, Yuan Y, Zheng W (2009) Adaptive workload prediction of grid performance in confidence windows. IEEE Trans Parallel Distrib Syst 21(7):925–938

    Google Scholar 

  3. Priya V, Babu CNK (2017) Moving average fuzzy resource scheduling for virtualized cloud data services. Comput Stand Interfaces 50:251–257

    Article  Google Scholar 

  4. Fu X, Zhou C (2017) Predicted affinity based virtual machine placement in cloud computing environments. IEEE Trans Cloud Comput 8(1):246–255

    Article  MathSciNet  Google Scholar 

  5. Di S, Kondo D, Cirne W (2014) Google hostload prediction based on Bayesian model with optimized feature combination. J Parallel Distrib Comput 74(1):1820–1832

    Article  Google Scholar 

  6. Islam S, Keung J, Lee K, Liu A (2012) Empirical prediction models for adaptive resource provisioning in the cloud. Futur Gener Comput Syst 28(1):155–162

    Article  Google Scholar 

  7. Hu R, Jiang J, Liu G, Wang L et al (2014) Efficient resources provisioning based on load forecasting in cloud. Sci World J

  8. Lu Y, Panneerselvam J, Liu L, Wu Y et al (2016) Rvlbpnn: a workload forecasting model for smart cloud computing. Sci Programm

  9. Graves A, Graves A (2012) Long short-term memory. Supervised sequence labelling with recurrent neural networks, 37–45

  10. Song B, Yu Y, Zhou Y, Wang Z, Du S (2018) Host load prediction with long short-term memory in cloud computing. J Supercomput 74:6554–6568

    Article  Google Scholar 

  11. Kumar J, Goomer R, Singh AK (2018) Long short term memory recurrent neural network (lstm-rnn) based workload forecasting model for cloud datacenters. Procedia Comput Sci 125:676–682

    Article  Google Scholar 

  12. Patel E, Kushwaha DS (2022) A hybrid cnn-lstm model for predicting server load in cloud computing. J Supercomput 78(8):1–30

    Article  Google Scholar 

  13. Minh TN, Nam T, Epema DH (2013) Parallel workload modeling with realistic characteristics. IEEE Trans Parallel Distrib Syst 25(8):2138–2148

    Article  Google Scholar 

  14. Cortez E, Bonde A, Muzio A, Russinovich M, Fontoura M, Bianchini R (2017) Resource central: Understanding and predicting workloads for improved resource management in large cloud platforms. In: Proceedings of the 26th Symposium on Operating Systems Principles, pp 153–167

  15. Wilkes J (2020) Yet more Google compute cluster trace data. Google research blog, Mountain View, CA, USA. Posted at https://ai.googleblog.com/2020/04/yet-more-google-compute-cluster-trace.html

  16. Dinda PA, O’Hallaron DR (2000) Host load prediction using linear models. Clust Comput 3:265–280

    Article  Google Scholar 

  17. Gong Z, Gu X, Wilkes J (2010) Press: Predictive elastic resource scaling for cloud systems. In: 2010 International Conference on Network and Service Management, pp 9–16. IEEE

  18. Khan A, Yan X, Tao S, Anerousis N (2012) Workload characterization and prediction in the cloud: A multiple time series approach. In: 2012 IEEE Network Operations and Management Symposium, pp. 1287–1294. IEEE

  19. Bey KB, Benhammadi F, Mokhtari A, Guessoum Z (2009) Cpu load prediction model for distributed computing. In: 2009 Eighth International Symposium on Parallel and Distributed Computing, pp. 39–45. IEEE

  20. Cao J, Fu J, Li M, Chen J (2014) Cpu load prediction for cloud environment based on a dynamic ensemble model. Softw Practice Exp 44(7):793–804

    Article  Google Scholar 

  21. Yang Q, Zhou Y, Yu Y, Yuan J, Xing X, Du S (2015) Multi-step-ahead host load prediction using autoencoder and echo state networks in cloud computing. J Supercomput 71:3037–3053

    Article  Google Scholar 

  22. Gupta S, Dileep AD, Gonsalves TA (2018) A joint feature selection framework for multivariate resource usage prediction in cloud servers using stability and prediction performance. J Supercomput 74:6033–6068

    Article  Google Scholar 

  23. Gupta S, Dileep AD, Gonsalves TA (2020) Online sparse blstm models for resource usage prediction in cloud datacentres. IEEE Trans Netw Serv Manage 17(4):2335–2349

    Article  Google Scholar 

  24. Karim ME, Maswood MMS, Das S, Alharbi AG (2021) Bhyprec: a novel bi-lstm based hybrid recurrent neural network model to predict the cpu workload of cloud virtual machine. IEEE Access 9:131476–131495

    Article  Google Scholar 

  25. Rossi A, Visentin A, Prestwich S, Brown KN (2022) Bayesian uncertainty modelling for cloud workload prediction. In: 2022 IEEE 15th International Conference on Cloud Computing (CLOUD), pp. 19–29. IEEE

  26. Saxena D, Kumar J, Singh AK, Schmid S (2023) Performance analysis of machine learning centered workload prediction models for cloud. IEEE Trans Parallel Distrib Syst 34(4):1313–1330

    Article  Google Scholar 

  27. Xu M, Song C, Wu H, Gill SS, Ye K, Xu C (2022) esdnn: deep neural network based multivariate workload prediction in cloud computing environments. ACM Trans Internet Technol (TOIT) 22(3):1–24

    Google Scholar 

  28. Cho K, Van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078

  29. Adebiyi AA, Adewumi AO, Ayo CK (2014) Comparison of arima and artificial neural networks models for stock price prediction. J Appl Math 2014(1):614342

    MathSciNet  Google Scholar 

  30. Masdari M, Khoshnevis A (2020) A survey and classification of the workload forecasting methods in cloud computing. Clust Comput 23(4):2399–2424. https://doi.org/10.1007/s10586-019-03010-3

    Article  Google Scholar 

  31. Dai W, Yang Q, Xue G-R, Yu Y (2007) Boosting for transfer learning. In: Proceedings of the 24th International Conference on Machine Learning, pp. 193–200

  32. Freund Y, Schapire RE et al (1996) Experiments with a new boosting algorithm. In: Icml, vol. 96, pp. 148–156. Citeseer

  33. Pardoe D, Stone P (2010) Boosting for regression transfer. In: Proceedings of the 27th International Conference on International Conference on Machine Learning, pp. 863–870

  34. Wang D, Yan M, Teng Y, Han D, Dang H, Ye X, Fan D (2023) A transfer learning framework for high-accurate cross-workload design space exploration of cpu. In: 2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD), pp. 1–9. IEEE

  35. Liu C, Jiao J, Li W, Wang J, Zhang J (2022) Tr-predictior: An ensemble transfer learning model for small-sample cloud workload prediction. Entropy 24(12):1770

    Article  Google Scholar 

  36. Wen Q, He K, Sun L, Zhang Y, Ke M, Xu H (2021) Robustperiod: Robust time-frequency mining for multiple periodicity detection. In: Proceedings of the 2021 International Conference on Management of Data, pp. 2328–2337

  37. Vlachos M, Yu PS, Castelli V, Meek C (2006) Structural periodic measures for time-series data. Data Min Knowl Disc 12:1–28

    Article  MathSciNet  Google Scholar 

  38. Chen F, Qin Z, Zhao H, Zhou M, Deng S (2023) Pepnet: A periodicity-perceived workload prediction network supporting rare occurrence of heavy workload. arXiv preprint arXiv:2308.01917

  39. Wu J, He J (2020) Continuous transfer learning with label-informed distribution alignment. arXiv preprint arXiv:2006.03230

  40. Guo Y, Shi H, Kumar A, Grauman K, Rosing T, Feris R (2019) Spottune: transfer learning through adaptive fine-tuning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4805–4814

  41. Yosinski J, Clune J, Bengio Y, Lipson H (2024) How transferable are features in deep neural networks? Adv Neural Inf Process Syst 27

  42. Long M, Cao Y, Cao Z, Wang J, Jordan MI (2018) Transferable representation learning with deep adaptation networks. IEEE Trans Pattern Anal Mach Intell 41(12):3071–3085

    Article  Google Scholar 

  43. Khanh NQ, Van T, Cuong D, Tuyen NL, Duc T (2024) Boosted Regression for Predicting CPU Utilization. https://github.com/khanhhust96/Boosted-Regression-for-Predicting-CPU-Utilization-in-the-Cloud-with-Periodicity Accessed on August 2024

  44. Wen L, Xu M, Toosi AN, Ye K (2024) Temposcale: A cloud workloads prediction approach integrating short-term and long-term information. In: 2024 IEEE 17th International Conference on Cloud Computing (CLOUD)

Download references

Acknowledgments

This research is funded by Hanoi University of Science and Technology (HUST) under the project number T2022-PC-048.

Author information

Authors and Affiliations

Authors

Contributions

All authors participated in the writing and review of the manuscript. Van Tong and Duc Tran proposed ideas for the manuscript. Khanh Nguyen Quoc and Cuong Dao generated all the figures and tables.

Corresponding author

Correspondence to Duc Tran.

Ethics declarations

Conflict of interest

The authors declare no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Quoc, K.N., Tong, V., Dao, C. et al. Boosted regression for predicting CPU utilization in the cloud with periodicity. J Supercomput 80, 26036–26060 (2024). https://doi.org/10.1007/s11227-024-06451-9

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-024-06451-9

Keywords