Abstract
Predicting CPU usage is crucial to cloud resource management. Precise CPU prediction, however, is a tough challenge due to the variable and dynamic nature of CPUs. In this paper, we introduce TrAdaBoost.WLP, a novel regression transfer boosting method that employs Long Short-Term Memory (LSTM) networks for CPU consumption prediction. Concretely, a dedicated Periodicity-aware LSTM (PA-LSTM) model is specifically developed to take into account the use of periodically repeated patterns in time series data while making predictions. To adjust for variations in CPU demands, multiple PA-LSTMs are trained and concatenated in TrAdaBoost.WLP using a boosting mechanism. TrAdaBoost.WLP and benchmarks have been thoroughly evaluated on two datasets: 160 Microsoft Azure VMs and 8 Google cluster traces. The experimental results show that TrAdaBoost.WLP can produce promising performance, improving by 32.4% and 59.3% in terms of mean squared error compared to the standard Probabilistic LSTM and ARIMA.








Similar content being viewed by others
References
Dogani J, Khunjush F, Seydali M (2023) Host load prediction in cloud computing with discrete wavelet transformation (dwt) and bidirectional gated recurrent unit (bigru) network. Comput Commun 198:157–174
Wu Y, Hwang K, Yuan Y, Zheng W (2009) Adaptive workload prediction of grid performance in confidence windows. IEEE Trans Parallel Distrib Syst 21(7):925–938
Priya V, Babu CNK (2017) Moving average fuzzy resource scheduling for virtualized cloud data services. Comput Stand Interfaces 50:251–257
Fu X, Zhou C (2017) Predicted affinity based virtual machine placement in cloud computing environments. IEEE Trans Cloud Comput 8(1):246–255
Di S, Kondo D, Cirne W (2014) Google hostload prediction based on Bayesian model with optimized feature combination. J Parallel Distrib Comput 74(1):1820–1832
Islam S, Keung J, Lee K, Liu A (2012) Empirical prediction models for adaptive resource provisioning in the cloud. Futur Gener Comput Syst 28(1):155–162
Hu R, Jiang J, Liu G, Wang L et al (2014) Efficient resources provisioning based on load forecasting in cloud. Sci World J
Lu Y, Panneerselvam J, Liu L, Wu Y et al (2016) Rvlbpnn: a workload forecasting model for smart cloud computing. Sci Programm
Graves A, Graves A (2012) Long short-term memory. Supervised sequence labelling with recurrent neural networks, 37–45
Song B, Yu Y, Zhou Y, Wang Z, Du S (2018) Host load prediction with long short-term memory in cloud computing. J Supercomput 74:6554–6568
Kumar J, Goomer R, Singh AK (2018) Long short term memory recurrent neural network (lstm-rnn) based workload forecasting model for cloud datacenters. Procedia Comput Sci 125:676–682
Patel E, Kushwaha DS (2022) A hybrid cnn-lstm model for predicting server load in cloud computing. J Supercomput 78(8):1–30
Minh TN, Nam T, Epema DH (2013) Parallel workload modeling with realistic characteristics. IEEE Trans Parallel Distrib Syst 25(8):2138–2148
Cortez E, Bonde A, Muzio A, Russinovich M, Fontoura M, Bianchini R (2017) Resource central: Understanding and predicting workloads for improved resource management in large cloud platforms. In: Proceedings of the 26th Symposium on Operating Systems Principles, pp 153–167
Wilkes J (2020) Yet more Google compute cluster trace data. Google research blog, Mountain View, CA, USA. Posted at https://ai.googleblog.com/2020/04/yet-more-google-compute-cluster-trace.html
Dinda PA, O’Hallaron DR (2000) Host load prediction using linear models. Clust Comput 3:265–280
Gong Z, Gu X, Wilkes J (2010) Press: Predictive elastic resource scaling for cloud systems. In: 2010 International Conference on Network and Service Management, pp 9–16. IEEE
Khan A, Yan X, Tao S, Anerousis N (2012) Workload characterization and prediction in the cloud: A multiple time series approach. In: 2012 IEEE Network Operations and Management Symposium, pp. 1287–1294. IEEE
Bey KB, Benhammadi F, Mokhtari A, Guessoum Z (2009) Cpu load prediction model for distributed computing. In: 2009 Eighth International Symposium on Parallel and Distributed Computing, pp. 39–45. IEEE
Cao J, Fu J, Li M, Chen J (2014) Cpu load prediction for cloud environment based on a dynamic ensemble model. Softw Practice Exp 44(7):793–804
Yang Q, Zhou Y, Yu Y, Yuan J, Xing X, Du S (2015) Multi-step-ahead host load prediction using autoencoder and echo state networks in cloud computing. J Supercomput 71:3037–3053
Gupta S, Dileep AD, Gonsalves TA (2018) A joint feature selection framework for multivariate resource usage prediction in cloud servers using stability and prediction performance. J Supercomput 74:6033–6068
Gupta S, Dileep AD, Gonsalves TA (2020) Online sparse blstm models for resource usage prediction in cloud datacentres. IEEE Trans Netw Serv Manage 17(4):2335–2349
Karim ME, Maswood MMS, Das S, Alharbi AG (2021) Bhyprec: a novel bi-lstm based hybrid recurrent neural network model to predict the cpu workload of cloud virtual machine. IEEE Access 9:131476–131495
Rossi A, Visentin A, Prestwich S, Brown KN (2022) Bayesian uncertainty modelling for cloud workload prediction. In: 2022 IEEE 15th International Conference on Cloud Computing (CLOUD), pp. 19–29. IEEE
Saxena D, Kumar J, Singh AK, Schmid S (2023) Performance analysis of machine learning centered workload prediction models for cloud. IEEE Trans Parallel Distrib Syst 34(4):1313–1330
Xu M, Song C, Wu H, Gill SS, Ye K, Xu C (2022) esdnn: deep neural network based multivariate workload prediction in cloud computing environments. ACM Trans Internet Technol (TOIT) 22(3):1–24
Cho K, Van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078
Adebiyi AA, Adewumi AO, Ayo CK (2014) Comparison of arima and artificial neural networks models for stock price prediction. J Appl Math 2014(1):614342
Masdari M, Khoshnevis A (2020) A survey and classification of the workload forecasting methods in cloud computing. Clust Comput 23(4):2399–2424. https://doi.org/10.1007/s10586-019-03010-3
Dai W, Yang Q, Xue G-R, Yu Y (2007) Boosting for transfer learning. In: Proceedings of the 24th International Conference on Machine Learning, pp. 193–200
Freund Y, Schapire RE et al (1996) Experiments with a new boosting algorithm. In: Icml, vol. 96, pp. 148–156. Citeseer
Pardoe D, Stone P (2010) Boosting for regression transfer. In: Proceedings of the 27th International Conference on International Conference on Machine Learning, pp. 863–870
Wang D, Yan M, Teng Y, Han D, Dang H, Ye X, Fan D (2023) A transfer learning framework for high-accurate cross-workload design space exploration of cpu. In: 2023 IEEE/ACM International Conference on Computer Aided Design (ICCAD), pp. 1–9. IEEE
Liu C, Jiao J, Li W, Wang J, Zhang J (2022) Tr-predictior: An ensemble transfer learning model for small-sample cloud workload prediction. Entropy 24(12):1770
Wen Q, He K, Sun L, Zhang Y, Ke M, Xu H (2021) Robustperiod: Robust time-frequency mining for multiple periodicity detection. In: Proceedings of the 2021 International Conference on Management of Data, pp. 2328–2337
Vlachos M, Yu PS, Castelli V, Meek C (2006) Structural periodic measures for time-series data. Data Min Knowl Disc 12:1–28
Chen F, Qin Z, Zhao H, Zhou M, Deng S (2023) Pepnet: A periodicity-perceived workload prediction network supporting rare occurrence of heavy workload. arXiv preprint arXiv:2308.01917
Wu J, He J (2020) Continuous transfer learning with label-informed distribution alignment. arXiv preprint arXiv:2006.03230
Guo Y, Shi H, Kumar A, Grauman K, Rosing T, Feris R (2019) Spottune: transfer learning through adaptive fine-tuning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4805–4814
Yosinski J, Clune J, Bengio Y, Lipson H (2024) How transferable are features in deep neural networks? Adv Neural Inf Process Syst 27
Long M, Cao Y, Cao Z, Wang J, Jordan MI (2018) Transferable representation learning with deep adaptation networks. IEEE Trans Pattern Anal Mach Intell 41(12):3071–3085
Khanh NQ, Van T, Cuong D, Tuyen NL, Duc T (2024) Boosted Regression for Predicting CPU Utilization. https://github.com/khanhhust96/Boosted-Regression-for-Predicting-CPU-Utilization-in-the-Cloud-with-Periodicity Accessed on August 2024
Wen L, Xu M, Toosi AN, Ye K (2024) Temposcale: A cloud workloads prediction approach integrating short-term and long-term information. In: 2024 IEEE 17th International Conference on Cloud Computing (CLOUD)
Acknowledgments
This research is funded by Hanoi University of Science and Technology (HUST) under the project number T2022-PC-048.
Author information
Authors and Affiliations
Contributions
All authors participated in the writing and review of the manuscript. Van Tong and Duc Tran proposed ideas for the manuscript. Khanh Nguyen Quoc and Cuong Dao generated all the figures and tables.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Quoc, K.N., Tong, V., Dao, C. et al. Boosted regression for predicting CPU utilization in the cloud with periodicity. J Supercomput 80, 26036–26060 (2024). https://doi.org/10.1007/s11227-024-06451-9
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-024-06451-9