Abstract:
Deep convolutional neural networks are hindered from empowering resource-constrained devices due to their demanding computational and storage resources. Structured prunin...Show MoreMetadata
Abstract:
Deep convolutional neural networks are hindered from empowering resource-constrained devices due to their demanding computational and storage resources. Structured pruning effectively removes the redundant components from neural networks and obtains compact models. Previous pruning methods usually evaluate the importance of filters from a layer-wise perspective, which is deprived of global guidance. We propose an adaptive pruning algorithm based on relevance scores to evaluate the contribution of each channel by calculating its relevance score from the back-propagation of the neural network's output. Our method identifies and removes channels with low contribution from a global perspective. Unlike previous methods that manually set the pruning rate for each pruning iteration, our method adaptively adjusts the pruning rate. In addition, our method performs satisfactorily with limited data for one-shot pruning in the absence of fine-tuning. The ability to obtain compact models through one-shot pruning with limited data is ideally suited for edge computing scenarios. We validate the effectiveness of our method with multiple combinations of convolutional neural networks and datasets. Our approach outperforms existing pruning methods in scenarios with limited data.
Date of Conference: 28 May 2023 - 01 June 2023
Date Added to IEEE Xplore: 23 October 2023
ISBN Information:
Electronic ISSN: 1938-1883