Abstract
Complex resource usage patterns of scaling Cloud workloads and heterogeneous infrastructure remain a challenge for accurate modelling of server load, which is the key to effective capacity sizing and provisioning in data centers. Recently, Long Short-Term Memory (LSTM) network has been used for host load prediction. However, learning complex noisy variations in host load is still an issue that needs to be addressed. In this work, we propose pCNN-LSTM, a hybrid prediction approach comprising of 1-dimensional Convolution Neural Networks (1D CNN) and LSTM, to predict CPU utilization on Cloud servers at multiple consecutive time-steps. It consists of three parallel dilated 1D CNN layers with different dilation rates for pattern extraction from noisy host CPU usage and an LSTM layer that learns temporal dependencies within the raw usage values as well as within the patterns extracted by the 1D CNN layers. Convolutions with different dilation rates enable the model to learn CPU load variations at different scales. Prediction skill of pCNN-LSTM is demonstrated using Google cluster trace, Alibaba trace and Bitbrains data, and performance is measured using Mean Squared Error (MSE) and Root Mean Squared Error (RMSE). pCNN-LSTM achieves up to 15%, 13% and 16% improvements in host load prediction with Google Trace, Alibaba trace and Bitbrains data set, respectively, over LSTM, Bidirectional LSTM (BLSTM), CNN-LSTM, CNN-BLSTM and two of its variants, showing the effectiveness of multi-scale learning capability of pCNN-LSTM and establishes its applicability as an adaptive prediction method for improved capacity planning and provisioning.















Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.References
Guo J, Chang Z, Wang S, Ding H, Feng Y, Mao L, Bao Y (2019). Who limits the resource efficiency of my datacenter: an analysis of alibaba datacenter traces. In: 2019 IEEE/ACM 27th international symposium on quality of service (IWQoS), pp 1–10. IEEE
Verma A, Pedrosa L, Korupolu M, Oppenheimer D, Tune E, Wilkes J (2015) Large-scale cluster management at Google with Borg. In: Proceedings of the Tenth European Conference on Computer Systems, pp 1–17
Tirmazi M, Barker A, Deng N, Haque M. E, Qin ZG, Hand S, Wilkes J et al (2020). Borg: the next generation. In: Proceedings of the fifteenth European conference on computer systems, pp 1–14
Reiss C, Tumanov A, Ganger GR, Katz RH, Kozuch MA (2012) Heterogeneity and dynamicity of clouds at scale: Google trace analysis. In: Proceedings of the third ACM symposium on cloud computing, pp 1–13
Varghese B, Buyya R (2018) Next generation cloud computing: new trends and research directions. Future Gener Comput Syst 79:849–861
Deng L, Ren YL, Xu F, He H, Li C (2018) Resource utilization analysis of Alibaba cloud. In: International conference on Intelligent Computing, pp 183–194. Springer, Cham
Makridakis S, Spiliotis E, Assimakopoulos V (2018) Statistical and Machine Learning forecasting methods: concerns and ways forward. PloS one 13(3):e0194889
Borovykh A, Bohte S, Oosterlee CW (2018) Dilated convolutional neural networks for time series forecasting. J Comput Financ, Forthcoming
Nemirovsky D, Arkose T, Markovic N, Nemirovsky M, Unsal O, Cristal A, Valero M (2018) A general guide to applying machine learning to computer architecture. Supercomput Front Innov 5(1):95–115
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Graves A, Liwicki M, Fernández S, Bertolami R, Bunke H, Schmidhuber J (2008) A novel connectionist system for unconstrained handwriting recognition. IEEE Trans Pattern Anal Mach Intel 31(5):855–868
Beaufays F (2015) The neural networks behind Google Voice transcription. Google Research blog. https://ai.googleblog.com/2015/08/the-neural-networks-behind-google-voice.html Accessed 03 (June 2021)
Baccouche M, Mamalet F, Wolf C, Garcia C, Baskurt A (2011) Sequential deep learning for human action recognition. In: International workshop on human behavior understanding, pp 29–39. Springer, Berlin
Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, Dean J et al (2016) Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144
Zhao Z, Chen W, Wu X, Chen PC, Liu J (2017) LSTM network: a deep learning approach for short-term traffic forecast. IET Intel Transp Syst 11(2):68–75
Kiranyaz S, Avci O, Abdeljaber O, Ince T, Gabbouj M, Inman DJ (2021) 1D convolutional neural networks and applications: a survey. Mech Syst Signal Process 151:107398
Wang Z, Yan W, Oates T (2017) Time series classification from scratch with deep neural networks: a strong baseline. In: 2017 International joint conference on neural networks (IJCNN), pp 1578–1585. IEEE
Lim B, Zohren S (2021) Time-series forecasting with deep learning: a survey. Philos Trans R Soc A 379(2194):20200209
Gao J, Wang H, Shen H (2020) Machine Learning Based Workload Prediction in Cloud Computing. In: 2020 29th international conference on computer communications and networks (ICCCN) pp 1-9. IEEE
Chen J, Wang Y (2019) A hybrid method for short-term host utilization prediction in cloud computing. J Electr Comput Eng (2019)
Janardhanan D, Barrett E (2017) CPU workload forecasting of machines in data centers using LSTM recurrent neural networks and ARIMA models. In: 2017 12th international conference for internet technology and secured transactions (ICITST), pp 55–60. IEEE
Duggan M, Mason K, Duggan J, Howley E, Barrett E (2017) Predicting host CPU utilization in cloud computing using recurrent neural networks. In: 2017 12th international conference for internet technology and secured transactions (ICITST), pp 67–72. IEEE
Yang Q, Zhou Y, Yu Y, Yuan J, Xing X, Du S (2015) Multi-step-ahead host load prediction using autoencoder and echo state networks in cloud computing. J Supercomput 71(8):3037–3053
Song B, Yu Y, Zhou Y, Wang Z, Du S (2018) Host load prediction with long short-term memory in cloud computing. J Supercomput 74(12):6554–6568
Tang X (2019) Large-scale computing systems workload prediction using parallel improved LSTM neural network. IEEE Access 7:40525–40533
Gupta S, Dinesh DA (2017) Resource usage prediction of cloud workloads using deep bidirectional long short term memory networks. In 2017 IEEE international conference on advanced networks and telecommunications systems (ANTS), pp 1–6. IEEE
Huang Z, Peng J, Lian H, Guo J, Qiu W (2017) Deep recurrent model for server load and performance prediction in data center. Complexity, 2017
Kumar J, Goomer R, Singh AK (2018) Long short term memory recurrent neural network (LSTM-RNN) based workload forecasting model for cloud datacenters. Proc Comput Sci 125:676–682
Tran N, Nguyen T, Nguyen BM, Nguyen G (2018) A multivariate fuzzy time series resource forecast model for clouds using LSTM and data correlation analysis. Proc Comput Sci 126:636–645
Gao J, Wang H, Shen H (2020) Task failure prediction in cloud data centers using deep learning. IEEE Trans Serv Comput
Karim ME, Maswood MMS, Das S, Alharbi AG (2021). BHyPreC: a novel Bi-LSTM based hybrid recurrent neural network model to predict the CPU workload of cloud virtual machine. IEEE Access
Wang K, Li K, Zhou L, Hu Y, Cheng Z, Liu J, Chen C (2019) Multiple convolutional neural networks for multivariate time series prediction. Neurocomputing 360:107–119
Kim TY, Cho SB (2019) Predicting residential energy consumption using CNN-LSTM neural networks. Energy 182:72–81
Le T, Vo MT, Vo B, Hwang E, Rho S, Baik SW (2019) Improving electric energy consumption prediction using CNN and Bi-LSTM. Appl Sci 9(20):4237
Qin D, Yu J, Zou G, Yong R, Zhao Q, Zhang B (2019) A novel combined prediction scheme based on CNN and LSTM for urban PM 2.5 concentration. IEEE Access 7:20050–20059
Ma L, Tian S (2020) A hybrid CNN-LSTM model for aircraft 4D trajectory prediction. IEEE Access 8:134668–134680
Wang H, Yang Z, Yu Q, Hong T, Lin X (2018) Online reliability time series prediction via convolutional neural network and long short term memory for service-oriented systems. Knowledge-Based Syst 159:132–147
Dumoulin V, Visin F (2016). A guide to convolution arithmetic for deep learning. arXiv preprint arXiv:1603.07285
Yu F, Koltun V (2015) Multi-scale context aggregation by dilated convolutions. arXiv preprint arXiv:1511.07122
Chollet F (2018) Deep learning with Python, vol 361. Manning, New York
Reiss C, Wilkes J, Hellerstein JL (2011) Google cluster-usage traces: format+ schema. Google Inc., White Paper, pp 1–14
Shen S, Van Beek V, Iosup A (2015). Statistical characterization of business-critical workloads hosted in cloud datacenters. In: 2015 15th IEEE/ACM international symposium on cluster, cloud and grid computing, pp 465–474. IEEE
Di S, Kondo D, Cirne W (2012) Host load prediction in a Google compute cloud with a Bayesian model. In: SC’12: proceedings of the international conference on high performance computing, networking, storage and analysis, pp 1–11. IEEE
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Patel, E., Kushwaha, D.S. A hybrid CNN-LSTM model for predicting server load in cloud computing. J Supercomput 78, 1–30 (2022). https://doi.org/10.1007/s11227-021-04234-0
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11227-021-04234-0