skip to main content
10.1145/3234804.3234824acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicdltConference Proceedingsconference-collections
research-article

Improvement of Pruning Method for Convolution Neural Network Compression

Authors Info & Claims
Published:27 June 2018Publication History

ABSTRACT

The large number of parameters in convolutional neural network (CNN) makes it a computationally intensive and storage-intensive network model. Although the effect of CNN is prominent in various identification and classification tasks, it is difficult to deploy on embedded devices because the model is too large. In order to solve this problem, an improved scheme for pruning operations in compression methods is proposed. First, the distribution of network connection is analyzed so as to determine the pruning threshold initially; then, using the pruning method to delete connections whose weights are less than the threshold, make the network quickly reach the limit of pruning but maintain accuracy. The verification experiment was performed on the Lenet-5 network which trained on the MINST data set and Lenet-5 was compressed 10.56 times without loss of accuracy.

References

  1. A. Krizhevsky, I. Sutskever, and G. E Hinton,"Imagenet classification with deepconvolutional neural networks". In Advances in neural information processing systems, pp.1097--1105, 2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. D. Amodei, R. Anubhai, E. Battenberg, et al.,"Deep speech 2: End-to-end speech recognition in english and mandarin". arXiv preprint arXiv:1512.02595, 2015.Google ScholarGoogle Scholar
  3. M. Luong, H. Pham, and C. D Manning. "Effective approaches to attention-based neural machine translation". arXiv preprint arXiv:1508.04025, 2015.Google ScholarGoogle Scholar
  4. M. Horowitz. "Energy table for 45nm process", Stanford VLSI wiki.Google ScholarGoogle Scholar
  5. S. Han, H. Mao, W. J. Dally, "Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding", ICLR'16.Google ScholarGoogle Scholar
  6. F. Iandola, S. Han, M. Moskewicz, et al. "SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size". 2016.Google ScholarGoogle Scholar
  7. M. Rastegari, V. Ordonez, J. Redmon, et al. "XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks". European Conference on Computer Vision, pp. 525--542, 2016.Google ScholarGoogle ScholarCross RefCross Ref
  8. A. G Howard, M. Zhu, B. Chen, et al. "MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications". arXiv preprint arXiv:1704.0486, 2017.Google ScholarGoogle Scholar
  9. Z. Xiang, Z. Xin, L. Meng, et al. "ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices". arXiv preprint arXiv:1707.01083, 2017.Google ScholarGoogle Scholar
  10. S. Han, J. Pool, J.Tran, and W. Dally. "Learning both weights and connections for efficient neural network". In Advances in Neural Information Processing Systems, pp.1135--1143, 2015. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. "Gradient-based learningapplied to document recognition". Proceedings of the IEEE, vol.86, pp.2278--2324, 1998.Google ScholarGoogle ScholarCross RefCross Ref
  12. https://stacks.stanford.edu/file/druid:qf934gh3708/EFFICIENT%20METHODS%20AND%20HARDWARE%20FOR%20DEEP%20LEARNING-augmented.pdfGoogle ScholarGoogle Scholar
  13. K. Simonyan and A. Zisserman. "Very deep convolutional networks for large-scale image recognition". arXiv preprint arXiv:1409.1556, 2014.Google ScholarGoogle Scholar
  14. C. Szegedy, W. Liu, Y. Jia, et al. "Going deeper with convolutions".In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.1--9, 2015.Google ScholarGoogle ScholarCross RefCross Ref
  15. K. He, X. Zhang, S. Ren, and J. Sun. "Deep residual learning for imagerecognition". arXiv preprint arXiv:1512.03385, 2015.Google ScholarGoogle Scholar
  16. Y. Han, T. Jiang T, Y. Ma, et al."Compression of deep neural networks", Application Research of Computers, vol. 35, pp. 1--7, 2017.Google ScholarGoogle Scholar
  17. Y. Jia, et al. "Caffe: Convolutional architecture for fast feature embedding". arXiv preprint arXiv:1408.5093, 2014.Google ScholarGoogle Scholar

Index Terms

  1. Improvement of Pruning Method for Convolution Neural Network Compression

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICDLT '18: Proceedings of the 2018 2nd International Conference on Deep Learning Technologies
      June 2018
      112 pages
      ISBN:9781450364737
      DOI:10.1145/3234804

      Copyright © 2018 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 27 June 2018

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader