Abstract
In general, deep neural networks (DNNs) are seriously overparameterized with enormous hardware resources demanded, which creates a heavy burden for inference applications especially for resource-constrained edge devices. To overcome this difficulty, there are two principal solutions: optimizing the overparameterized DNNs and designing high-efficiency DNN algorithms with lightweight architectures. In terms of the optimization methods, pruning is the most effective technique because the solution fundamentally optimizes the bloated DNNs by removing redundant structures from the network and can be seamlessly incorporated into all other optimization solutions as well as all kinds of DNN architectures. Nevertheless, the study reported in this paper reveals that the various excellent but also complicated pruning algorithms may not be as effective as proposals demonstrate and do not yield optimal solutions for all cases. In addition, the current lightweight DNN architectures are also overparameterized to a large extent. In this research, we propose a mechanism for determining lightweight DNN networks From Scratch (FS-DNN). First, we conduct a thorough study on the theoretical basis of evaluating the hardware resources demanded by DNNs, and establish the objective function for determining a lightweight DNN network. Based on the study, the theoretical FS-DNN for determining lightweight DNNs from scratch with high efficiency is proposed. Then, we perform a series of experiments with FS-DNN based lightweight DNNs on the public dataset CIFAR10/100 and a private dataset Kuzushiji, which prove the feasibility and efficiency of FS-DNN. According to the research, instead of adopting bloated DNN networks that demand complicated pruning algorithms to optimize the networks after the fact or the current so-called lightweight DNNs, the experimental results demonstrate that lightweight networks based on FS-DNN achieve superior performance in computing consumption with competitive or even better accuracy.


Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
All data generated or analysed during this study is included in this published article (and its supplementary information files).
References
Yue X, Li H, Shimizu M, Kawamura S, Meng L (2022) Yolo-gd: a deep learning-based object detection algorithm for empty-dish recycling robots. Machines
Kim J, Kim Han-Joon, Kim H (2019) Fraud detection for job placement using hierarchical clusters-based deep neural networks. Appl Intell 49(8):2842–2861
Yue X, Li H, Fujikawa Y, Meng L (2022) Dynamic dataset augmentation for deep learning-based oracle bone inscriptions recognition. ACM Journal on Computing and Cultural Heritage (JOCCH)
Tiwari S, Goel S, Bhardwaj A (2022) Midnn- a classification approach for the eeg based motor imagery tasks using deep neural network. Appl Intell 52(5):4824–4843
Ji D, Shin D, Park J (2021) An error compensation technique for low-voltage dnn accelerators. IEEE Trans Very Large Scale Integr (VLSI) Syst 29(2):397–408
Trappey AJC, Trappey CV, Govindarajan UH, Sun JJH (2021) Patent value analysis using deep learning models—the case of iot technology mining for the manufacturing industry. IEEE Trans Eng Manag 68 (5):1334–1346
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: Advances in neural information processing systems, vol 25. Curran Associates, Inc, Lake Tahoe, Nevada, USA
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: The 3rd international conference on learning representations(ICLR), San diego, CA, USA
Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: 2015 IEEE Conference on computer vision and pattern recognition (CVPR), Boston, MA, USA, pp 1–9
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: The 32nd international conference on machine learning (ICML), vol 37, Lille, France, pp 448–456
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: 2016 IEEE Conference on computer vision and pattern recognition (CVPR), Las Vegas, NV, USA, pp 2818–2826
He K, Zhang X, Ren S, Sun J (June 2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on computer vision and pattern recognition (CVPR), Los Alamitos, CA, USA, pp 770–778
Huang G, Liu Z, Van Der Maaten L, Weinberger KQ (2017) Densely connected convolutional networks. In: 2017 IEEE Conference on computer vision and pattern recognition (CVPR), Los Alamitos, CA, USA, pp 2261–2269
Zhang X, Zhou X, Lin M, Sun J (2018) Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: 2018 IEEE Conference on computer vision and pattern recognition (CVPR), Salt Lake City, pp 6848–6856
Ma N et al (2018) Shufflenet v2: practical guidelines for efficient CNN architecture design. In: The 15th european conference on computer vision (ECCV), vol 11218, Munich, Germany, pp 122–138
Tan M, Bo C, Pang R, Vasudevan V, Le QV (2019) Mnasnet: platform-aware neural architecture search for mobile. In: 2019 IEEE/CVF Conference on computer vision and pattern recognition (CVPR), Long Beach, CA, USA, pp 2815–2823
Sandler M, Howard A, Zhu M, Zhmoginov A, Chen Liang-Chieh (2018) Mobilenetv2: inverted residuals and linear bottlenecks. In: 2018 IEEE/CVF Conference on computer vision and pattern recognition, pp 4510–4520
Li H, Wang Z, Yue X, Wang W, Tomiyama H, Meng L (2022) An architecture-level analysis on deep learning models for low-impact computations. Artif Intell Rev, 55
Liu W, Wu G, Ren F, Kang X (2020) Dff-resnet: an insect pest recognition model based on residual networks. Big Data Mining and Analytics 3(4):300–310
Yu X, Kang C, Guttery DS, Kadry S, Chen Y, Zhang Y-D (2021) Resnet-scda-50 for breast abnormality classification. IEEE/ACM Trans Comput Biol Bioinform 18(1):94–102
Lin C, Zhao G, Yang Z, Yin A, Wang X, Li G, Chen H, Ma Z, Zhao L, Luo H, Wang T, Ding B, Pang X, Chen Q (2022) Cir-net: automatic classification of human chromosome based on inception-resnet architecture. IEEE/ACM Trans Comput Biol Bioinform 19(3):1285–1293
Huang L, Pun C-M (2020) Audio replay spoof attack detection by joint segment-based linear filter bank feature extraction and attention-enhanced densenet-bilstm network. IEEE/ACM Trans Audio Speech Lang Process 28:1813–1825
Bakshi S, Rajan S (2021) Fall event detection system using inception-densenet inspired sparse siamese network. IEEE Sensors Lett 5(8):1–4
Yang X, Ye Q, Cai G, Wang Y, Cai G (2022) Pd-resnet for classification of parkinson’s disease from gait. IEEE J Transl Eng Health Med 10:1–11
Li H, Yue X, Wang Z, Chai Z, Wang W, Tomiyama H, Meng L (2022) Optimizing the deep neural networks by layer-wise refined pruning and the acceleration on fpga. Computational Intelligence and Neuroscience, 2022
Li T, Wu B, Yang Y, Fan Y, Zhang Y, Liu W (2019) Compressing convolutional neural networks via factorized convolutional filters. In: 2019 IEEE/CVF conference on computer vision and pattern recognition (CVPR), Long Beach, CA, USA, pp 3972–3981
Zhang Q, Han R, Xin G, Liu CH, Wang G, Chen LY (2022) Lightweight and accurate dnn-based anomaly detection at edge. IEEE Trans Parallel Distrib Syst 33(11):2927–2942
Li B, Qu S, Wang Y (2022) An automated quantization framework for high-utilization rram-based pim. IEEE Trans Comput-Aided Des Integr Circuits Syst 41(3):583–596
Choi Y, El-Khamy M, Lee J (2020) Learning sparse low-precision neural networks with learnable regularization. IEEE Access 8:96963–96974
Gong C, Chen Y, Ye L u, Li T, Hao C, Chen D (2021) Vecq: minimal loss dnn model compression with vectorized weight quantization. IEEE Trans Comput 70(5):696–710
Kim N, Shin D, Choi W, Kim G, Park J (2021) Exploiting retraining-based mixed-precision quantization for low-cost dnn accelerator design. IEEE Trans Neural Netw Learn Syst 32(7):2925–2938
Su Y, Hong D, Li Y, Jing P (2020) Low-rank regularized deep collaborative matrix factorization for micro-video multi-label classification. IEEE Signal Process Lett 27:740–744
Xue Z, Du J, Du D, Li G, Huang Q, Lyu S (2019) Deep constrained low-rank subspace learning for multi-view semi-supervised classification. IEEE Signal Process Lett 26(8):1177–1181
Liao Z, Huang J, Cheng Y, Li C, Liu PX (2022) A novel decomposition-based ensemble model for short-term load forecasting using hybrid artificial neural networks. Appl Intell 52(10):11043–11057
Yazdani R, Riera M, Arnau J-M, González A (2018) The dark side of dnn pruning. In: 2018 ACM/IEEE 45Th annual international symposium on computer architecture (ISCA), pp 790–801
Fang Y u, Li C, Wang P, Han C, Huang R, Xi H (2021) Easiedge: a novel global deep neural networks pruning method for efficient edge computing. IEEE Internet Things J 8(3):1259–1271
Zhang T, Ye S, Feng X, Ma X, Zhang K, Li Z, Tang J, Liu S, Lin X, Liu Y, Fardad M, Wang Y (2022) Structadmm: achieving ultrahigh efficiency in structured pruning for dnns. IEEE Trans Neural Netw Learn Syst 33(5):2259–2273
Zhuang T, Zhang Z, Huang Y, Zeng X, Shuang K, Li X (2020) Neuron-level structured pruning using polarization regularizer. In: Larochelle H, Ranzato M, Hadsell R, Balcan MF, Lin H (eds) Advances in Neural Information Processing Systems, vol 33. Curran Associates, Inc., pp 9865–9877
Chen T, Bo J i, Tianyu DING, Fang B, Wang G, Zhu Z, Liang L, Shi Y, Yi S, Xiao T u (2021) Only train once: a one-shot neural network training and pruning framework. In: Thirty-fifth conference on neural information processing systems
Loshchilov I, Hutter F (2017) SGDR: Stochastic gradient descent with warm restarts. In: The 5th international conference on learning representations (ICLR), Toulon, France
He Y, Liu P, Wang Z, Hu Z, Yi Y (2019) Filter pruning via geometric median for deep convolutional neural networks acceleration. In: 2019 IEEE/CVF Conference on computer vision and pattern recognition (CVPR), Long Beach, CA, USA, 06, pp 4335–4344
Huang Z, Li L i, Hailin Sun. (2020) Global biased pruning considering layer contribution. IEEE Access 8:173521–173529
Center for Open Data in the Humanities. Kuzushiji dataset (in japanese). 2018. http://codh.rois.ac.jp/char-shape/book/100249416/2019,. (Apr. 22, 2021 accessed)
Meng L Character recognition, 2022-08 accessed
Huang Y, Qiao X, Ren P, Liu L, Pu C, Dustdar S, Chen J (2022) A lightweight collaborative deep neural network for the mobile web in edge cloud. IEEE Trans Mob Comput 21(7):2289– 2305
Geng T, Li A, Wang T, Wu C, Li Y, Shi R, Wu W, Herbordt M (2021) O3bnn-r: an out-of-order architecture for high-performance and regularized bnn inference. IEEE Transactions on Parallel and Distributed Systems 32(1):199–213
Pham T-N, Trinh Q-K, Chang I-J, Alioto M (2022) Stt-bnn: a novel stt-mram in-memory computing macro for binary neural networks. IEEE Journal on Emerging and Selected Topics in Circuits and Systems 12 (2):569–579
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix A: Block structures of Lightweight-DNNs
Appendix A: Block structures of Lightweight-DNNs
- MobileNetV2: :
-
Figure 2 shows two kinds of MobileNet-bottlenecks. The bottleneck adopts depthwise convolution (DWConv) for the [3 × 3] convolution to reduce the model size and complexity. The [1 × 1] convolution is used to increase/decrease the dimensions.
- ShuffleNet Block: :
-
Figure 3 shows the ShuffleNet-block structures. In addition to taking depthwise convolution (DWConv) to adjust the dimensions, the ShuffleNet block has two branches, which are concatenated and then the channels are shuffled.
ShuffleNet Block [15]
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Li, H., Yue, X., Zhao, C. et al. Lightweight deep neural network from scratch. Appl Intell 53, 18868–18886 (2023). https://doi.org/10.1007/s10489-022-04394-3
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-04394-3