skip to main content
10.1145/3583780.3614889acmconferencesArticle/Chapter ViewAbstractPublication PagescikmConference Proceedingsconference-collections
research-article

FVW: Finding Valuable Weight on Deep Neural Network for Model Pruning

Published:21 October 2023Publication History

ABSTRACT

The rapid development of deep learning has demonstrated its potential for deployment in many intelligent service systems. However, some issues such as optimisation (e.g., how to reduce the deployment resources costs and further improve the detection speed), especially in scenarios where limited resources are available, remain challenging to address. In this paper, we aim to delve into the principles of deep neural networks, focusing on the importance of network neurons. The goal is to identify the neurons that exert minimal impact on model performances, thereby aiding in the process of model pruning. In this work, we have thoroughly considered the deep learning model pruning process with and without fine-tuning step, ensuring the model performance consistency. To achieve our objectives, we propose a methodology that employs adversarial attack methods to explore deep neural network parameters. This approach is combined with an innovative attribution algorithm to analyse the level of network neurons involvement. In our experiments, our approach can effectively quantify the importance of network neuron. We extend the evaluation through comprehensive experiments conducted on a range of datasets, including CIFAR-10, CIFAR-100 and Caltech101. The results demonstrate that, our method have consistently achieved the state-of-the-art performance over many existing methods. We anticipate that this work will help to reduce the heavy training and inference cost of deep neural network models where a lightweight deep learning enhanced service and system is possible. The source code is open source at https://github.com/LMBTough/FVW.

References

  1. A. Voulodimos, N. Doulamis, A. Doulamis, E. Protopapadakis et al., "Deep learning for computer vision: A brief review," Computational intelligence and neuroscience, vol. 2018, 2018.Google ScholarGoogle Scholar
  2. A. Osipov, E. Pleshakova, S. Gataullin, S. Korchagin, M. Ivanov, A. Finogeev, and V. Yadav, "Deep learning method for recognition and classification of images from video recorders in difficult weather conditions," Sustainability, vol. 14, no. 4, p. 2420, 2022.Google ScholarGoogle ScholarCross RefCross Ref
  3. S. Mehtab and J. Sen, "Analysis and forecasting of financial time series using cnn and lstm-based deep learning models," in Advances in Distributed Computing and Machine Learning: Proceedings of ICADCML 2021. Springer, 2022, pp. 405--423.Google ScholarGoogle Scholar
  4. M. E. Alzahrani, T. H. Aldhyani, S. N. Alsubari, M. M. Althobaiti, and A. Fahad, "Developing an intelligent system with deep learning algorithms for sentiment analysis of e-commerce product reviews," Computational Intelligence and Neuro-science, vol. 2022, 2022.Google ScholarGoogle Scholar
  5. S. Han, J. Pool, J. Tran, and W. Dally, "Learning both weights and connections for efficient neural network," in Advances in Neural Information Processing Systems, C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett, Eds., vol. 28. Curran Associates, Inc., 2015.Google ScholarGoogle Scholar
  6. M. Uzair and N. Jamil, "Effects of hidden layers on the efficiency of neural networks," in 2020 IEEE 23rd international multitopic conference (INMIC). IEEE, 2020, pp. 1--6.Google ScholarGoogle Scholar
  7. Y. Y. Huang and W. Y. Wang, "Deep residual learning for weakly-supervised relation extraction," arXiv preprint arXiv:1707.08866, 2017.Google ScholarGoogle Scholar
  8. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, L. Kaiser, and I. Polosukhin, "Attention is all you need," Advances in neural information processing systems, vol. 30, 2017.Google ScholarGoogle Scholar
  9. M. H. Zhu and S. Gupta, "To prune, or not to prune: Exploring the efficacy of pruning for model compression," 2018. [Online]. Available: https://openreview.net/forum"id=S1lN69AT-Google ScholarGoogle Scholar
  10. P. Molchanov, A. Mallya, S. Tyree, I. Frosio, and J. Kautz, "Importance estimation for neural network pruning," in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 11 264--11 272.Google ScholarGoogle Scholar
  11. H. Hu, R. Peng, Y.-W. Tai, and C.-K. Tang, "Network trimming: A data-driven neuron pruning approach towards efficient deep architectures," arXiv preprint arXiv:1607.03250, 2016.Google ScholarGoogle Scholar
  12. T.-J. Yang, Y.-H. Chen, and V. Sze, "Designing energy-efficient convolutional neural networks using energy-aware pruning," in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 5687--5695.Google ScholarGoogle Scholar
  13. Y. LeCun, J. Denker, and S. Solla, "Optimal brain damage," Advances in neural information processing systems, vol. 2, 1989.Google ScholarGoogle Scholar
  14. M. Mondal, B. Das, S. D. Roy, P. Singh, B. Lall, and S. D. Joshi, "Adaptive cnn filter pruning using global importance metric," Computer Vision and Image Understanding, vol. 222, p. 103511, 2022.Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. S. Anwar, K. Hwang, and W. Sung, "Structured pruning of deep convolutional neural networks," ACM Journal on Emerging Technologies in Computing Systems (JETC), vol. 13, no. 3, pp. 1--18, 2017.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. C. Zhao, B. Ni, J. Zhang, Q. Zhao, W. Zhang, and Q. Tian, "Variational convolutional neural network pruning," in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 2780--2789.Google ScholarGoogle Scholar
  17. Z. Wang, C. Li, and X. Wang, "Convolutional neural network pruning with structural redundancy reduction," in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 14 913--14 922.Google ScholarGoogle Scholar
  18. M. Sundararajan, A. Taly, and Q. Yan, "Axiomatic attribution for deep networks," in International conference on machine learning. PMLR, 2017, pp. 3319--3328.Google ScholarGoogle Scholar
  19. A. Patra and J. A. Noble, "Incremental learning of fetal heart anatomies using interpretable saliency maps," in Medical Image Understanding and Analysis: 23rd Conference, MIUA 2019, Liverpool, UK, July 24--26, 2019, Proceedings 23. Springer, 2020, pp. 129--141.Google ScholarGoogle Scholar
  20. Z. Wang, M. Fredrikson, and A. Datta, "Robust models are more interpretable because attributions look normal," arXiv preprint arXiv:2103.11257, 2021.Google ScholarGoogle Scholar
  21. I. J. Goodfellow, J. Shlens, and C. Szegedy, "Explaining and harnessing adversarial examples," arXiv preprint arXiv:1412.6572, 2014.Google ScholarGoogle Scholar
  22. A. Kurakin, I. J. Goodfellow, and S. Bengio, "Adversarial examples in the physical world," in Artificial intelligence safety and security. Chapman and Hall/CRC, 2018, pp. 99--112.Google ScholarGoogle Scholar
  23. Y. Dong, F. Liao, T. Pang, H. Su, J. Zhu, X. Hu, and J. Li, "Boosting adversarial attacks with momentum," in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 9185--9193.Google ScholarGoogle Scholar
  24. J. Lin, C. Song, K. He, L. Wang, and J. E. Hopcroft, "Nesterov accelerated gradient and scale invariance for adversarial attacks," arXiv preprint arXiv:1908.06281, 2019.Google ScholarGoogle Scholar
  25. A. Krizhevsky, V. Nair, and G. Hinton, "The cifar-10 dataset," online: http://www.cs. toronto.edu/kriz/cifar. html, vol. 55, no. 5, 2014.Google ScholarGoogle Scholar
  26. --, "Cifar-10 and cifar-100 datasets," URl: https://www. cs. toronto. edu/kriz/cifar.html, vol. 6, no. 1, p. 1, 2009.Google ScholarGoogle Scholar
  27. F.-F. Li, M. Andreeto, M. Ranzato, and P. Perona, "Caltech 101," 4 2022.Google ScholarGoogle Scholar
  28. K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770--778.Google ScholarGoogle Scholar
  29. H. Wang, C. Qin, Y. Zhang, and Y. Fu, "Neural pruning via growing regularization," arXiv preprint arXiv:2012.09243, 2020.Google ScholarGoogle Scholar
  30. G. Retsinas, A. Elafrou, G. Goumas, and P. Maragos, "Weight pruning via adaptive sparsity loss," arXiv preprint arXiv:2006.02768, 2020.Google ScholarGoogle Scholar

Index Terms

  1. FVW: Finding Valuable Weight on Deep Neural Network for Model Pruning

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Conferences
        CIKM '23: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management
        October 2023
        5508 pages
        ISBN:9798400701245
        DOI:10.1145/3583780

        Copyright © 2023 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 21 October 2023

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate1,861of8,427submissions,22%

        Upcoming Conference

      • Article Metrics

        • Downloads (Last 12 months)88
        • Downloads (Last 6 weeks)11

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader