skip to main content
10.1145/2934583.2934625acmconferencesArticle/Chapter ViewAbstractPublication PagesislpedConference Proceedingsconference-collections
research-article

Speeding up Convolutional Neural Network Training with Dynamic Precision Scaling and Flexible Multiplier-Accumulator

Published:08 August 2016Publication History

ABSTRACT

Training convolutional neural network is a major bottleneck when developing a new neural network topology. This paper presents a dynamic precision scaling (DPS) algorithm and flexible multiplier-accumulator (MAC) to speed up convolutional neural network training. The DPS algorithm utilizes dynamic fixed point and finds good enough numerical precision for target network while training. The precision information from DPS is used to configure our proposed MAC. The proposed MAC can perform fixed point computation with variable precision mode providing differentiated computation time which enables speeding up training for lower precision computation. Simulation results show that our work can achieve 5.7x speed-up while consuming 31% energy compared to baseline for modified Alexnet on Flickr image style recognition task.

References

  1. J. Dean et al., "Large scale distributed deep networks," Advances in Neural Information Processing Systems, pp. 1223--1231, 2012.Google ScholarGoogle Scholar
  2. T. Chilimbi et al., "Project Adam: Building an efficient and scalable deep learning training system," OSDI, pp. 571--582, October 2014 Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. A. Krizhevsky, I. Sutskever, and G. Hinton, "ImageNet Classification with Deep Convolutional Neural Networks," Proc. Neural Information and Processing Systems, 2012.Google ScholarGoogle Scholar
  4. C. Zhang et al., "Optimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks," FPGA'2015, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. J. Cong and B. Xiao. "Minimizing computation in convolutional neural networks," ICANN, pages 281--290. Springer, 2014.Google ScholarGoogle Scholar
  6. Z. Lin et al., "Neural Networks with Few Multiplications," arXiv preprint arXiv:1510.03009v3, 2016Google ScholarGoogle Scholar
  7. S. Gupta et al., "Deep learning with limited numerical precision," ICML, 2015Google ScholarGoogle Scholar
  8. M. Courbariaux et al., "Training Deep Neural Networks with Low Precision Multiplications," arXiv preprint arXiv:1412.7024, 2014.Google ScholarGoogle Scholar
  9. D. Williamson, "Dynamically scaled fixed point arithmetic," in proc. IEEE Conf. Commun., Comput. Syst. Signal Process., May 1991, pp. 315--318.Google ScholarGoogle Scholar
  10. D. Lin et al., "Fixed Point Quantization of Deep Convolutional Networks," arXiv preprint arXiv:1511.06393v2, 2016.Google ScholarGoogle Scholar
  11. Y. Lecun et al., "Gradient-based learning applied to document recognition," Proceedings of the IEEE, 86, no. 112278-2324, 1998.Google ScholarGoogle Scholar
  12. C. Szegedy et al., "Going Deeper with Convolutions," CVPR, 2015Google ScholarGoogle Scholar
  13. Y. Jia et al., "Caffe: Convolutional Architecture for Fast Feature Embedding," ACM International Conference on Multimedia, 2014. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. S. Chetlur et al., "cudnn: Efficient primitives for deep learning". CoRR, abs/1410.0759, 2014.Google ScholarGoogle Scholar
  15. H.T. Kung, "Why systolic architectures?," Computer, 15(1):37--46, Jan 1982. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Speeding up Convolutional Neural Network Training with Dynamic Precision Scaling and Flexible Multiplier-Accumulator

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          ISLPED '16: Proceedings of the 2016 International Symposium on Low Power Electronics and Design
          August 2016
          392 pages
          ISBN:9781450341851
          DOI:10.1145/2934583

          Copyright © 2016 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 8 August 2016

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article
          • Research
          • Refereed limited

          Acceptance Rates

          ISLPED '16 Paper Acceptance Rate60of190submissions,32%Overall Acceptance Rate398of1,159submissions,34%

          Upcoming Conference

          ISLPED '24

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader