Skip to main content
Log in

A hybrid CNN-LSTM model for predicting server load in cloud computing

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

Complex resource usage patterns of scaling Cloud workloads and heterogeneous infrastructure remain a challenge for accurate modelling of server load, which is the key to effective capacity sizing and provisioning in data centers. Recently, Long Short-Term Memory (LSTM) network has been used for host load prediction. However, learning complex noisy variations in host load is still an issue that needs to be addressed. In this work, we propose pCNN-LSTM, a hybrid prediction approach comprising of 1-dimensional Convolution Neural Networks (1D CNN) and LSTM, to predict CPU utilization on Cloud servers at multiple consecutive time-steps. It consists of three parallel dilated 1D CNN layers with different dilation rates for pattern extraction from noisy host CPU usage and an LSTM layer that learns temporal dependencies within the raw usage values as well as within the patterns extracted by the 1D CNN layers. Convolutions with different dilation rates enable the model to learn CPU load variations at different scales. Prediction skill of pCNN-LSTM is demonstrated using Google cluster trace, Alibaba trace and Bitbrains data, and performance is measured using Mean Squared Error (MSE) and Root Mean Squared Error (RMSE). pCNN-LSTM achieves up to 15%, 13% and 16% improvements in host load prediction with Google Trace, Alibaba trace and Bitbrains data set, respectively, over LSTM, Bidirectional LSTM (BLSTM), CNN-LSTM, CNN-BLSTM and two of its variants, showing the effectiveness of multi-scale learning capability of pCNN-LSTM and establishes its applicability as an adaptive prediction method for improved capacity planning and provisioning.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

References

  1. Guo J, Chang Z, Wang S, Ding H, Feng Y, Mao L, Bao Y (2019). Who limits the resource efficiency of my datacenter: an analysis of alibaba datacenter traces. In: 2019 IEEE/ACM 27th international symposium on quality of service (IWQoS), pp 1–10. IEEE

  2. Verma A, Pedrosa L, Korupolu M, Oppenheimer D, Tune E, Wilkes J (2015) Large-scale cluster management at Google with Borg. In: Proceedings of the Tenth European Conference on Computer Systems, pp 1–17

  3. Tirmazi M, Barker A, Deng N, Haque M. E, Qin ZG, Hand S, Wilkes J et al (2020). Borg: the next generation. In: Proceedings of the fifteenth European conference on computer systems, pp 1–14

  4. Reiss C, Tumanov A, Ganger GR, Katz RH, Kozuch MA (2012) Heterogeneity and dynamicity of clouds at scale: Google trace analysis. In: Proceedings of the third ACM symposium on cloud computing, pp 1–13

  5. Varghese B, Buyya R (2018) Next generation cloud computing: new trends and research directions. Future Gener Comput Syst 79:849–861

    Article  Google Scholar 

  6. Deng L, Ren YL, Xu F, He H, Li C (2018) Resource utilization analysis of Alibaba cloud. In: International conference on Intelligent Computing, pp 183–194. Springer, Cham

  7. Makridakis S, Spiliotis E, Assimakopoulos V (2018) Statistical and Machine Learning forecasting methods: concerns and ways forward. PloS one 13(3):e0194889

    Article  Google Scholar 

  8. Borovykh A, Bohte S, Oosterlee CW (2018) Dilated convolutional neural networks for time series forecasting. J Comput Financ, Forthcoming

  9. Nemirovsky D, Arkose T, Markovic N, Nemirovsky M, Unsal O, Cristal A, Valero M (2018) A general guide to applying machine learning to computer architecture. Supercomput Front Innov 5(1):95–115

    Google Scholar 

  10. Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

    Article  Google Scholar 

  11. Graves A, Liwicki M, Fernández S, Bertolami R, Bunke H, Schmidhuber J (2008) A novel connectionist system for unconstrained handwriting recognition. IEEE Trans Pattern Anal Mach Intel 31(5):855–868

    Article  Google Scholar 

  12. Beaufays F (2015) The neural networks behind Google Voice transcription. Google Research blog. https://ai.googleblog.com/2015/08/the-neural-networks-behind-google-voice.html Accessed 03 (June 2021)

  13. Baccouche M, Mamalet F, Wolf C, Garcia C, Baskurt A (2011) Sequential deep learning for human action recognition. In: International workshop on human behavior understanding, pp 29–39. Springer, Berlin

  14. Wu Y, Schuster M, Chen Z, Le QV, Norouzi M, Macherey W, Dean J et al (2016) Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv preprint arXiv:1609.08144

  15. Zhao Z, Chen W, Wu X, Chen PC, Liu J (2017) LSTM network: a deep learning approach for short-term traffic forecast. IET Intel Transp Syst 11(2):68–75

    Article  Google Scholar 

  16. Kiranyaz S, Avci O, Abdeljaber O, Ince T, Gabbouj M, Inman DJ (2021) 1D convolutional neural networks and applications: a survey. Mech Syst Signal Process 151:107398

    Article  Google Scholar 

  17. Wang Z, Yan W, Oates T (2017) Time series classification from scratch with deep neural networks: a strong baseline. In: 2017 International joint conference on neural networks (IJCNN), pp 1578–1585. IEEE

  18. Lim B, Zohren S (2021) Time-series forecasting with deep learning: a survey. Philos Trans R Soc A 379(2194):20200209

    Article  MathSciNet  Google Scholar 

  19. Gao J, Wang H, Shen H (2020) Machine Learning Based Workload Prediction in Cloud Computing. In: 2020 29th international conference on computer communications and networks (ICCCN) pp 1-9. IEEE

  20. Chen J, Wang Y (2019) A hybrid method for short-term host utilization prediction in cloud computing. J Electr Comput Eng (2019)

  21. Janardhanan D, Barrett E (2017) CPU workload forecasting of machines in data centers using LSTM recurrent neural networks and ARIMA models. In: 2017 12th international conference for internet technology and secured transactions (ICITST), pp 55–60. IEEE

  22. Duggan M, Mason K, Duggan J, Howley E, Barrett E (2017) Predicting host CPU utilization in cloud computing using recurrent neural networks. In: 2017 12th international conference for internet technology and secured transactions (ICITST), pp 67–72. IEEE

  23. Yang Q, Zhou Y, Yu Y, Yuan J, Xing X, Du S (2015) Multi-step-ahead host load prediction using autoencoder and echo state networks in cloud computing. J Supercomput 71(8):3037–3053

    Article  Google Scholar 

  24. Song B, Yu Y, Zhou Y, Wang Z, Du S (2018) Host load prediction with long short-term memory in cloud computing. J Supercomput 74(12):6554–6568

    Article  Google Scholar 

  25. Tang X (2019) Large-scale computing systems workload prediction using parallel improved LSTM neural network. IEEE Access 7:40525–40533

    Article  Google Scholar 

  26. Gupta S, Dinesh DA (2017) Resource usage prediction of cloud workloads using deep bidirectional long short term memory networks. In 2017 IEEE international conference on advanced networks and telecommunications systems (ANTS), pp 1–6. IEEE

  27. Huang Z, Peng J, Lian H, Guo J, Qiu W (2017) Deep recurrent model for server load and performance prediction in data center. Complexity, 2017

  28. Kumar J, Goomer R, Singh AK (2018) Long short term memory recurrent neural network (LSTM-RNN) based workload forecasting model for cloud datacenters. Proc Comput Sci 125:676–682

    Article  Google Scholar 

  29. Tran N, Nguyen T, Nguyen BM, Nguyen G (2018) A multivariate fuzzy time series resource forecast model for clouds using LSTM and data correlation analysis. Proc Comput Sci 126:636–645

    Article  Google Scholar 

  30. Gao J, Wang H, Shen H (2020) Task failure prediction in cloud data centers using deep learning. IEEE Trans Serv Comput

  31. Karim ME, Maswood MMS, Das S, Alharbi AG (2021). BHyPreC: a novel Bi-LSTM based hybrid recurrent neural network model to predict the CPU workload of cloud virtual machine. IEEE Access

  32. Wang K, Li K, Zhou L, Hu Y, Cheng Z, Liu J, Chen C (2019) Multiple convolutional neural networks for multivariate time series prediction. Neurocomputing 360:107–119

    Article  Google Scholar 

  33. Kim TY, Cho SB (2019) Predicting residential energy consumption using CNN-LSTM neural networks. Energy 182:72–81

    Article  Google Scholar 

  34. Le T, Vo MT, Vo B, Hwang E, Rho S, Baik SW (2019) Improving electric energy consumption prediction using CNN and Bi-LSTM. Appl Sci 9(20):4237

    Article  Google Scholar 

  35. Qin D, Yu J, Zou G, Yong R, Zhao Q, Zhang B (2019) A novel combined prediction scheme based on CNN and LSTM for urban PM 2.5 concentration. IEEE Access 7:20050–20059

    Article  Google Scholar 

  36. Ma L, Tian S (2020) A hybrid CNN-LSTM model for aircraft 4D trajectory prediction. IEEE Access 8:134668–134680

    Article  Google Scholar 

  37. Wang H, Yang Z, Yu Q, Hong T, Lin X (2018) Online reliability time series prediction via convolutional neural network and long short term memory for service-oriented systems. Knowledge-Based Syst 159:132–147

    Article  Google Scholar 

  38. Dumoulin V, Visin F (2016). A guide to convolution arithmetic for deep learning. arXiv preprint arXiv:1603.07285

  39. Yu F, Koltun V (2015) Multi-scale context aggregation by dilated convolutions. arXiv preprint arXiv:1511.07122

  40. Chollet F (2018) Deep learning with Python, vol 361. Manning, New York

    Google Scholar 

  41. Reiss C, Wilkes J, Hellerstein JL (2011) Google cluster-usage traces: format+ schema. Google Inc., White Paper, pp 1–14

    Google Scholar 

  42. Shen S, Van Beek V, Iosup A (2015). Statistical characterization of business-critical workloads hosted in cloud datacenters. In: 2015 15th IEEE/ACM international symposium on cluster, cloud and grid computing, pp 465–474. IEEE

  43. Di S, Kondo D, Cirne W (2012) Host load prediction in a Google compute cloud with a Bayesian model. In: SC’12: proceedings of the international conference on high performance computing, networking, storage and analysis, pp 1–11. IEEE

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Eva Patel.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Patel, E., Kushwaha, D.S. A hybrid CNN-LSTM model for predicting server load in cloud computing. J Supercomput 78, 1–30 (2022). https://doi.org/10.1007/s11227-021-04234-0

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-021-04234-0

Keywords

Navigation