Skip to main content

Evaluating Performance, Power and Energy of Deep Neural Networks on CPUs and GPUs

  • Conference paper
  • First Online:
Theoretical Computer Science (NCTCS 2021)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1494))

Included in the following conference series:

Abstract

Deep learning has achieved accuracy and fast training speed and has been successfully applied to many fields, including speech recognition, text processing, image processing and video processing. However, the cost of high power and energy comes together with the high accuracy and training speed of Deep Neural Network (DNN). This inspires researchers to perform characterization in terms of performance, power and energy for guiding the architecture design of DNN models. There are three critical issues to solve for designing a both accurate and energy-efficient DNN model: i) how the software parameters affect the DNN models; ii) how the hardware parameters affect the DNN models; and iii) how to choose the best energy-efficient DNN model. To answer the three issues above, we capture and analyze the performance, power and energy behaviors for multiple experiment settings. We evaluate four DNN models (i.e., LeNet, GoogLeNet, AlexNet, and CaffeNet) with various parameter settings (both hardware and software) on both CPU and GPU platforms. Evaluation results provide detailed DNN characterization and some key insights to facilitate the design of energy-efficient deep learning solutions.

Y. Sun and Z. Ou contributed equally to this work and should be considered as co-first authors.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Convnet. https://github.com/soumith/convnet-benchmarks

  2. cxxnet. https://github.com/dmlc/cxxnet

  3. Imagenet large-scale visual recognition challenge. http://image-net.org/challenges/LSVRC

  4. Nvidia system management interface. https://developer.nvidia.com/nvidia-system-management-interface

  5. Abadi, M., et al.: TensorFlow: large-scale machine learning on heterogeneous distributed systems. ArXiv abs/1603.04467 (2016)

    Google Scholar 

  6. Cai, E., Juan, D.C., Stamoulis, D., Marculescu, D.: NeuralPower: predict and deploy energy-efficient convolutional neural networks. In: The 9th Asian Conference on Machine Learning (ACML 2017) (2017)

    Google Scholar 

  7. Chen, J., et al.: Analyzing time-dimension communication characterizations for representative scientific applications on supercomputer systems. Front. Comput. Sci. 13(6), 1228–1242 (2019)

    Article  Google Scholar 

  8. Chen, T., et al.: MXNet: a flexible and efficient machine learning library for heterogeneous distributed systems. ArXiv abs/1512.01274 (2015)

    Google Scholar 

  9. Collobert, R., Kavukcuoglu, K., Farabet, C.: Torch7: a matlab-like environment for machine learning. In: NIPS 2011 (2011)

    Google Scholar 

  10. Committe, G.: Green500. https://www.top500.org/lists/green500/. Accessed 20 May 2021

  11. Guassic: Text classification with CNN and RNN. https://github.com/gaussic/text-classification-cnn-rnn

  12. Han, S., Pool, J., Tran, J., Dally, W.J.: Learning both weights and connections for efficient neural networks. In: Twenty-ninth Conference on Neural Information Processing Systems (NIPS 2015) (2015)

    Google Scholar 

  13. Hodak, M., Gorkovenko, M., Dholakia, A.: Towards power efficiency in deep learning on data center hardware. In: 2019 IEEE International Conference on Big Data (Big Data) (2019)

    Google Scholar 

  14. Jia, Y., et al.: Caffe: convolutional architecture for fast feature embedding. ArXiv abs/1408.5093 (2014)

    Google Scholar 

  15. Khan, K.N., Hirki, M., Niemi, T., Nurminen, J.K., Ou, Z.: RAPL in action: experiences in using RAPL for power measurements. ACM Trans. Model. Perform. Eval. Comput. Syst. 3(2), 1–26 (2018)

    Article  Google Scholar 

  16. Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25(2), 1097–1105 (2012)

    Google Scholar 

  17. LeCun, Y., Boser, B., Denker, J.S., Henderson, D., Howard, R.E.: Backpropagation applied to handwritten zip code recognition. Neural Comput. 1(4), 541–551 (1989)

    Article  Google Scholar 

  18. LeCun, Y., Kavukcuoglu, K., Farabet, C.: Convolutional networks and applications in vision. In: Proceedings of 2010 IEEE International Symposium on Circuits and Systems, pp. 253–256 (2010). https://doi.org/10.1109/ISCAS.2010.5537907

  19. Li, D., Chen, X., Becchi, M., Zong, Z.: Evaluating the energy efficiency of deep convolutional neural networks on CPUs and GPUs. In: 2016 IEEE International Conferences on Big Data and Cloud Computing (BDCloud), Social Computing and Networking (SocialCom), Sustainable Computing and Communications (SustainCom) (BDCloud-SocialCom-SustainCom), pp. 477–484 (2016)

    Google Scholar 

  20. Sun, M., Li, J., Guo, Z.: THUCTC: an efficient Chinese text classification toolkit. http://thuctc.thunlp.org/

  21. Ou, Z., Chen, J., Zhang, Y., Dong, Y., Yuan, Y., Wang, Z.: Power modeling for Phytium FT-2000+/64 multi-core architecture. In PPoPP 2020 Workshop: Principles and Practice of Parallel Programming 2020, Workshop: Benchmarking in the Datacenter, 7 p. (2020)

    Google Scholar 

  22. Rodrigues, C.F., Riley, G., Luján, M.: Fine-grained energy profiling for deep convolutional neural networks on the Jetson TX1. In: 2017 IEEE International Symposium on Workload Characterization (IISWC), pp. 114–115 (2017)

    Google Scholar 

  23. Rouhani, B.D., Mirhoseini, A., Koushanfar, F.: DeLight: adding energy dimension to deep neural networks. In: International Symposium on Low Power Electronics & Design (2016)

    Google Scholar 

  24. Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Proceedings of the 27th International Conference on Neural Information Processing Systems, NIPS 2014, vol. 2, p. 3104–3112. MIT Press, Cambridge (2014)

    Google Scholar 

  25. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Rabinovich, A.: Going deeper with convolutions. IEEE Computer Society (2014)

    Google Scholar 

  26. Tang, Z., Wang, Y., Wang, Q., Chu, X.: The impact of GPU DVFs on the energy and performance of deep learning: an empirical study. In: The Tenth ACM International Conference (2019)

    Google Scholar 

  27. Thomas, D., Shanmugasundaram, M.: A survey on different overclocking methods. In: 2018 Second International Conference on Electronics, Communication and Aerospace Technology (ICECA), pp. 1588–1592 (2018)

    Google Scholar 

  28. Wu, F., et al.: A holistic energy-efficient approach for a processor-memory system. Tsinghua Sci. Technol. 24, 468–483 (2019)

    Article  Google Scholar 

  29. Yang, T.J., Chen, Y.H., Sze, V.: Designing energy-efficient convolutional neural networks using energy-aware pruning. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)

    Google Scholar 

  30. Yao, C., et al.: Evaluating and analyzing the energy efficiency of CNN inference on high-performance GPU. Pract. Exp. Concurr. Comput. 33, e6064 (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Juan Chen .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Sun, Y. et al. (2021). Evaluating Performance, Power and Energy of Deep Neural Networks on CPUs and GPUs. In: Cai, Z., Li, J., Zhang, J. (eds) Theoretical Computer Science. NCTCS 2021. Communications in Computer and Information Science, vol 1494. Springer, Singapore. https://doi.org/10.1007/978-981-16-7443-3_12

Download citation

  • DOI: https://doi.org/10.1007/978-981-16-7443-3_12

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-16-7442-6

  • Online ISBN: 978-981-16-7443-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics