Abstract
Channel pruning can reduce the number of neural network parameters and computational cost by eliminating redundant channels, its main purpose is to adapt to resource constrained devices. Evaluation-based global pruning and network search-based pruning are two common methods of channel pruning. However, the network architecture pruned by the global mask is often not optimal, while the method that directly searches for the optimal architecture will introduce a large number of hyperparameters, which greatly increases the training cost. In this paper, we propose a novel Two-dimensional information Entropy based Channel Pruning method (TECP). The pruning process consists of two steps. First, a global mask pruning scheme is employed to obtained a pre-pruning model. Then, the two-dimensional information entropy is calculated by using feature maps of dense network to adjust the pre-pruning model adaptively to get a compact network. Moreover, the entropy values are used to determine the minimum number of reserved channels per layer based on to avoid the imbalance of network architecture and the layer collapse caused by global pruning. Extensive experiments with a variety of networks on several datasets clearly demonstrate the effectiveness of our proposed TECP method. For example, results show that on CIFAR-10, the compressed model achieves comparable accuracy to the original model, but with a significantly lower number of parameters (44.29% for ResNet-20 and 46.79% for VGG-16). This is beneficial for industrial deployment. And experimental results also show that TECP method obtain the better performance compared with state-of-the-art method.









Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
No data was used for the research described in the article.
References
Masana M, Liu X, Twardowski B et al (2023) Class-incremental learning: Survey and performance evaluation on image classification. IEEE Trans Pattern Anal Mach Intell 45:5513–5533. https://doi.org/10.1109/TPAMI.2022.3213473
Ahmad HM, Rahimi A (2022) Deep learning methods for object detection in smart manufacturing: A survey. J Manuf Syst 64:181–196. https://doi.org/10.1016/j.jmsy.2022.06.011
Aljabri M, AlGhamdi M (2022) A review on the use of deep learning for medical images segmentation. Neurocomputing 506:311–335. https://doi.org/10.1016/j.neucom.2022.07.070
Paymode AS, Malode VB (2022) Transfer learning for multi-crop leaf disease image classification using convolutional neural network vgg. Artif Intell Agric 6:23–33. https://doi.org/10.1016/j.aiia.2021.12.002
Hadipour-Rokni R, Askari Asli-Ardeh E, Jahanbakhshi A et al (2023) Intelligent detection of citrus fruit pests using machine vision system and convolutional neural network through transfer learning technique. Comput Biol Med 155:106611. https://doi.org/10.1016/j.compbiomed.2023.106611
Mohanta BK, Jena D, Satapathy U et al (2020) Survey on iot security: Challenges and solution using machine learning, artificial intelligence and blockchain technology. Internet of Things 11:100227. https://doi.org/10.1016/j.iot.2020.100227
Yang J, Wang Y, Zhao H et al (2022) Mobilenet and knowledge distillation-based automatic scenario recognition method in vehicle-to-vehicle systems. IEEE Trans Veh Technol 71:11006–11016. https://doi.org/10.1109/TVT.2022.3184994
Yang H, Liu J, Mei G et al (2023) Research on real-time detection method of rail corrugation based on improved shufflenet v2. Eng Appl Artif Intell 126:106825. https://doi.org/10.1016/j.engappai.2023.106825
Liang T, Glossner J, Wang L et al (2021) Pruning and quantization for deep neural network acceleration: A survey. Neurocomputing 461:370–403. https://doi.org/10.1016/j.neucom.2021.07.045
Shuvo MMH, Islam SK, Cheng J et al (2023) Efficient acceleration of deep learning inference on resource-constrained edge devices: A review. Proc IEEE 111:42–91. https://doi.org/10.1109/JPROC.2022.3226481
Zhao R, Gui G, Xue Z et al (2022) A novel intrusion detection method based on lightweight neural network for internet of things. IEEE Internet of Things Journal 9:9960–9972. https://doi.org/10.1109/JIOT.2021.3119055
Chang J, Lu Y, Xue P et al (2022) Automatic channel pruning via clustering and swarm intelligence optimization for cnn. Appl Intell 52:17751–17771. https://doi.org/10.1007/s10489-022-03508-1
Peng J, Sun W, Li HC et al (2022) Low-rank and sparse representation for hyperspectral image processing: A review. IEEE Geosci Remote Sens Mag 10:10–43. https://doi.org/10.1109/MGRS.2021.3075491
Chu Y, Li P, Bai Y et al (2022) Group channel pruning and spatial attention distilling for object detection. Appl Intell 52:16246–16264. https://doi.org/10.1007/s10489-022-03293-x
Abdar M, Pourpanah F, Hussain S et al (2021) A review of uncertainty quantification in deep learning: Techniques, applications and challenges. Inf Fusion 76:243–297. https://doi.org/10.1016/j.inffus.2021.05.008
Tanaka H, Kunin D, Yamins DL, et al (2020) Pruning neural networks without any data by iteratively conserving synaptic flow. In: Advances in neural information processing systems, vol 33. Curran Associates, Inc., pp 6377–6389. https://proceedings.neurips.cc/paper_files/paper/2020/file/46a4378f835dc8040c8057beb6a2da52-Paper.pdf
Yeom SK, Seegerer P, Lapuschkin S et al (2021) Pruning by explaining: A novel criterion for deep neural network pruning. Pattern Recognit 115:107899. https://doi.org/10.1016/j.patcog.2021.107899
Tessier H, Gripon V, Léonardon M et al (2022) Rethinking weight decay for efficient neural network pruning. J Imaging 8:64. https://doi.org/10.3390/jimaging8030064
Wimmer P, Mehnert J, Condurache A (2022) Interspace pruning: Using adaptive filter representations to improve training of sparse cnns. In: 2022 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 12517–12527 https://doi.org/10.1109/CVPR52688.2022.01220
Abdelfattah A, Costa T, Dongarra J et al (2021) A set of batched basic linear algebra subprograms and lapack routines. ACM Trans Math Softw 47. https://doi.org/10.1145/3431921
Zhang Y, Freris NM (2023) Adaptive filter pruning via sensitivity feedback. IEEE Transactions on Neural Networks and Learning Systems, pp 1–13. https://doi.org/10.1109/TNNLS.2023.3246263
He Y, Liu P, Wang Z, et al (2019) Filter pruning via geometric median for deep convolutional neural networks acceleration. In: 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 4335–4344 https://doi.org/10.1109/CVPR.2019.00447
Zhu J, Pei J (2022) Progressive kernel pruning cnn compression method with an adjustable input channel. Appl Intell 52:1–22. https://doi.org/10.1007/s10489-021-02932-z
Yang W, Xiao Y (2022) Structured pruning via feature channels similarity and mutual learning for convolutional neural network compression. Appl Intell 52:14560–14570. https://doi.org/10.1007/s10489-022-03403-9
Wang Z, Li F, Shi G et al (2020) Network pruning using sparse learning and genetic algorithm. Neurocomputing 404:247–256. https://doi.org/10.1016/j.neucom.2020.03.082
Xie Y, Chen H, Ma Y et al (2022) Automated design of cnn architecture based on efficient evolutionary search. Neurocomputing 491:160–171. https://doi.org/10.1016/j.neucom.2022.03.046
He Y, Lin J, Liu Z, et al (2018) Amc: Automl for model compression and acceleration on mobile devices. In: Ferrari V, Hebert M, Sminchisescu C, et al (eds) Computer Vision – ECCV 2018. Springer International Publishing, Cham, pp 815–832 https://doi.org/10.1007/978-3-030-01234-2_48
Guo S, Wang Y, Li Q, et al (2020) Dmcp: Differentiable markov channel pruning for neural networks. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 1536–1544 https://doi.org/10.1109/CVPR42600.2020.00161
Wang L, Huang W, Zhang M et al (2022) Pruning graph neural networks by evaluating edge properties. Knowl-Based Syst 256:109847. https://doi.org/10.1016/j.knosys.2022.109847
Medhat S, Abdel-Galil H, Aboutabl AE, et al (2023) Iterative magnitude pruning-based light-version of alexnet for skin cancer classification. Neural Computing and Applications, pp 1–16. https://doi.org/10.1007/s00521-023-09111-w
Hou Z, Qin M, Sun F, et al (2022) Chex: Channel exploration for cnn model compression. In: 2022 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 12277–12288 https://doi.org/10.1109/CVPR52688.2022.01197
Zhang Z, Qi F, Liu Z et al (2021) Know what you don’t need: Single-shot meta-pruning for attention heads. AI Open 2:36–42. https://doi.org/10.1016/j.aiopen.2021.05.003
Hayashi T, Cimr D, Studnička F et al (2024) Distance-based one-class time-series classification approach using local cluster balance. Expert Syst Appl 235:121201. https://doi.org/10.1016/j.eswa.2023.121201
Luo JH, Wu J (2020) Neural network pruning with residual-connections and limited-data. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 1455–1464 https://doi.org/10.1109/CVPR42600.2020.00153
Tofigh S, Ahmad MO, Swamy M (2022) A low-complexity modified thinet algorithm for pruning convolutional neural networks. IEEE Signal Proc Lett 29:1012–1016. https://doi.org/10.1109/LSP.2022.3164328
He Y, Dong X, Kang G et al (2020) Asymptotic soft filter pruning for deep convolutional neural networks. IEEE Trans Cyberne 50:3594–3604. https://doi.org/10.1109/TCYB.2019.2933477
Huang GB, Mattar M, Berg T, et al (2008) Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In: Workshop on faces in’Real-Life’Images: detection, alignment, and recognition
Schroff F, Kalenichenko D, Philbin J (2015) Facenet: A unified embedding for face recognition and clustering. In: 2015 IEEE conference on computer vision and pattern recognition (CVPR), pp 815–823 https://doi.org/10.1109/CVPR.2015.7298682
Acknowledgements
This work was supported in part by the National Natural Science Foundation of China(NSFC) under Grant no.61973009.
Author information
Authors and Affiliations
Contributions
Yifei Xu: Investigation, Methodology, Writing- original draft. Jinfu Yang: Supervision, Writing - review & editing, Funding acquisition. Runshi Wang: Resources, Writing - review & editing. Haoqing Li: Validation, Writing - review & editing.
Corresponding author
Ethics declarations
Competing flicts of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Xu, Y., Yang, J., Wang, R. et al. An effective two-stage channel pruning method based on two-dimensional information entropy. Appl Intell 54, 8491–8504 (2024). https://doi.org/10.1007/s10489-024-05615-7
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-024-05615-7