Skip to main content

Deep Architecture Compression with Automatic Clustering of Similar Neurons

  • Conference paper
  • First Online:
Book cover Pattern Recognition and Computer Vision (PRCV 2021)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 13022))

Included in the following conference series:

  • 1710 Accesses

Abstract

The more complex the deep neural networks (DNNs) are, the more diverse the learning tasks they can be applied to. However, for complex DNNs, it is difficult to deploy them on to the edge devices, which have limited computation and storage resources. In this paper, we propose an automatic neurons clustering (ANC) approach for deep architecture compression, it can reduce the computation and storage consumption without degrading the model performance. Specifically, an automatic clustering algorithm is used to discover similar neurons in each layer of the deep architecture, then the similar neurons and the corresponding connections are merged based on the results of automatic clustering. After fine-tuning, a more compact and less storage space occupied neural network is obtained, with no performance degradation compared to the original deep architecture. This compression method is fully applicable to fully connected layer and convolutional layer, both of which are common modules of popular DNNs. The analysis of neuron redundancy in DNNs is performed on a deep belief network (DBN), and it is verified that there is great redundancy among neurons in DNNs. To verify the effectiveness of the proposed ANC, we conducted experiments on DBN and VGGNet-16 using MNIST, CIFAR-10 and CIFAR-100 datasets. The experimental results demonstrate that our method can effectively perform deep architecture compression without losing network performance. After fine-tuning, it can even obtain higher accuracy than the original network. In addition, the superiority of ANC is further demonstrated by comparing it with related network compression methods.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Chen, H., et al.: Data-free learning of student networks. In: 17 International Conference on Computer Vision, Seoul, pp. 3513–3521. IEEE Computer Society (2019)

    Google Scholar 

  2. Ding, X., Ding, G., Zhou, X., Guo, Y., Han, J., Liu, J.: Global sparse momentum SGD for pruning very deep neural networks. In: 32nd Annual Conference on Neural Information Processing Systems, Vancouver, pp. 6379–6391 (2019)

    Google Scholar 

  3. Ding, X., Zhang, X., Han, J., Ding, G.: Repmlp: re-parameterizing convolutions into fully-connected layers for image recognition. CoRR abs/2105.01883 (2021)

    Google Scholar 

  4. Ding, X., Zhang, X., Ma, N., Han, J., Ding, G., Sun, J.: Repvgg: making vgg-style convnets great again. CoRR abs/2101.03697 (2021)

    Google Scholar 

  5. of Electronic Engineering, D., Computer Science, U.o.M.: senmentation and clustering. [EB/OL] (2012). https://www.eecs.umich.edu/vision/teaching/EECS442_2012/lectures/seg_cluster.pdf

  6. Fukunaga, K., Hostetler, L.D.: The estimation of the gradient of a density function, with applications in pattern recognition. IEEE Trans. Inf. Theory 21(1), 32–40 (1975)

    Article  MathSciNet  Google Scholar 

  7. Goodfellow, I.J., et al.: Generative adversarial networks. CoRR abs/1406.2661 (2014)

    Google Scholar 

  8. Han, S., Mao, H., Dally, W.J.: Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. In: 4th International Conference on Learning Representations (ICLR), San Juan (2016)

    Google Scholar 

  9. Hassibi, B., Stork, D.G.: Second order derivatives for network pruning: Optimal brain surgeon. In: 5th Advances in Neural Information Processing Systems, Denver, Colorado, pp. 164–171. Morgan Kaufmann, Denver (1992)

    Google Scholar 

  10. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 4th IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, pp. 770–778. IEEE Computer Society (2016)

    Google Scholar 

  11. Hinton, G.E., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. CoRR abs/1503.02531 (2015)

    Google Scholar 

  12. Howard, A., et al.: Searching for mobilenetv3. In: 17th International Conference on Computer Vision, Seoul, pp. 1314–1324. IEEE Computer Society (2019)

    Google Scholar 

  13. Howard, A.G., et al.: Mobilenets: efficient convolutional neural networks for mobile vision applications. CoRR abs/1704.04861 (2017)

    Google Scholar 

  14. Huang, Y., et al.: Efficient training of giant neural networks using pipeline parallelism. In: 33rd Annual Conference on Neural Information Processing Systems (NIPS), Vancouver, pp. 103–112 (2019)

    Google Scholar 

  15. Iandola, F.N., Moskewicz, M.W., Ashraf, K., Han, S., Dally, W.J., Keutzer, K.: Squeezenet: Alexnet-level accuracy with 50x fewer parameters and<1mb model size. CoRR abs/1602.07360 (2016)

    Google Scholar 

  16. Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: 26th Advances in Neural Information Processing Systems, Lake Tahoe. pp. 1106–1114 (2012)

    Google Scholar 

  17. Lebedev, V., Lempitsky, V.S.: Fast convnets using group-wise brain damage. In: 4th IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, pp. 2554–2564. IEEE Computer Society (2016)

    Google Scholar 

  18. Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P.: Pruning filters for efficient convnets. In: 5th International Conference on Learning Representations (ICLR), Toulon (2017)

    Google Scholar 

  19. Liu, X., Wang, L., Liu, W., Zhong, G.: Incremental layers resection: a novel method to compress neural networks. IEEE Access 7, 172167–172177 (2019)

    Article  Google Scholar 

  20. Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., Zhang, C.: Learning efficient convolutional networks through network slimming. In: 16th International Conference on Computer Vision, Venice, pp. 2755–2763. IEEE Computer Society (2017)

    Google Scholar 

  21. Ma, N., Zhang, X., Zheng, H.-T., Sun, J.: ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision – ECCV 2018. LNCS, vol. 11218, pp. 122–138. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_8

    Chapter  Google Scholar 

  22. Romero, A., Ballas, N., Kahou, S.E., Chassang, A., Gatta, C., Bengio, Y.: Fitnets: hints for thin deep nets. In: 3rd International Conference on Learning Representations (ICLR), San Diego (2015)

    Google Scholar 

  23. Sandler, M., Howard, A.G., Zhu, M., Zhmoginov, A., Chen, L.: Mobilenetv 2: Inverted residuals and linear bottlenecks. In: 6th IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, pp. 4510–4520. IEEE Computer Society (2018)

    Google Scholar 

  24. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: 3rd International Conference on Learning Representations (ICLR), San Diego (2015)

    Google Scholar 

  25. Tolstikhin, I.O., et al.: Mlp-mixer: An all-mlp architecture for vision. CoRR abs/2105.01601 (2021)

    Google Scholar 

  26. Vaswani, A., et al.: Attention is all you need. In: 31st Annual Conference on Neural Information Processing Systems, Long Beach, pp. 5998–6008 (2017)

    Google Scholar 

  27. Wu, B., et al.: Visual transformers: token-based image representation and processing for computer vision. CoRR abs/2006.03677 (2020)

    Google Scholar 

  28. Zagoruyko, S., Komodakis, N.: Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. In: 5th International Conference on Learning Representations, Toulon. OpenReview.net (2017)

    Google Scholar 

  29. Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: 6th IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, pp. 6848–6856. IEEE Computer Society (2018)

    Google Scholar 

  30. Zhong, G., Liu, W., Yao, H., Li, T., Sun, J., Liu, X.: Merging similar neurons for deep networks compression. Cogn. Comput. 12(3), 577–588 (2020)

    Article  Google Scholar 

Download references

Acknowledgement

This work was supported by the National Key Research and Development Program of China under Grant No. 2018AAA0100400, the Joint Fund of the Equipments Pre-Research and Ministry of Education of China under Grant No. 6141A0 20337, the Science and Technology Program of Qingdao under Grant No. 21-1-4-ny-19-nsh, the Natural Science Foundation of Shandong Province under Grant No. ZR2020MF131, and the Open Fund of Engineering Research Center for Medical Data Mining and Application of Fujian Province under Grant No. MDM2018007. Thanks to Zhaoxu Ding for his assistance in writing this paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Guoqiang Zhong .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, X., Liu, W., Wang, LN., Zhong, G. (2021). Deep Architecture Compression with Automatic Clustering of Similar Neurons. In: Ma, H., et al. Pattern Recognition and Computer Vision. PRCV 2021. Lecture Notes in Computer Science(), vol 13022. Springer, Cham. https://doi.org/10.1007/978-3-030-88013-2_30

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-88013-2_30

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-88012-5

  • Online ISBN: 978-3-030-88013-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics