Abstract
To account for both network accuracy and deployment efficiency, multi-objective NAS has been proposed. However, these methods are still inefficient. As a result of the previous coarse-grained performance descriptions, the acquisition of feasible networks is limited, reducing the effectiveness and efficiency of search direction updates. As a consequence, a large number of ineffective network structures that do not meet the application requirements are sampled and trained, significantly reducing search efficiency. To address the two issues, this paper uses adaptive dataflow mapping to describe the inference latency of the sampled network structures in a fine-grained manner, thereby expanding the search space of available networks. The sampling behavior is dynamically adjusted in accordance with the system latency requirement and the allocable latency for each layer. The sampled network structure’s latency will approach the system requirement. Finally, by reducing the number of infeasible networks, search efficiency is improved. Experiments show that comparing state-of-the-art methods, the accuracy of the networks searched by our framework can be improved by up to 7.93\(\%\) while meeting specific inference latency requirements. However, search efficiency can be improved up to 2.35\(\times \).











Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.Data availability
The datasets generated during and/or analyzed during the current study are available from the corresponding author on reasonable request.
References
Liu H, Simonyan K, Vinyals O, C Fernando, Kavukcuoglu K (2018) Hierarchical representations for efficient architecture search. In: International Conference on Learning Representations (ICLR)
Real E, Moore S, Selle A, Saxena S, Suematsu YL, Tan J, Le QV, Kurakin A (2017) Large-scale evolution of image classifiers. In: International Conference on Machine Learning (ICML), 2902-2911
Zoph B, Le QV (2017) Neural architecture search with reinforcement learning. In: International Conference on Learning Representations (ICLR)
Liu Y, Sun Y, Xue B, Zhang M, Yen GG, Tan KC (2021) A survey on evolutionary neural architecture search. IEEE Trans Neural Netw Learn Syst 1–21
Ren P, Xiao Y, Chang X, Huang PY, Li Z, Chen X, Wang X (2021) A comprehensive survey of neural architecture search: challenges and solutions. ACM Comput Surv (CSUR) 54(4):1–34
Guerrero-Viu J, Hauns S, Izquierdo S, Miotto G, Schrodi S, Biedenkapp A, Elsken T, Deng D, Lindauer M, Hutter F (2021) Bag of baselines for multi-objective joint neural architecture search and hyperparameter optimization. In arXiv preprint arXiv:2105.01015
Pham H, Guan M, Zoph B, Le QV, Dean J (2018) Efficient neural architecture search via parameter sharing. In: International Conference on Machine Learning (ICML), 4095-4104
Real E, Aggarwal A, Huang Y, Le QV (2019) Regularized evolution for image classifier architecture search. In: Proceedings of the AAAI conference on artificial intelligence (AAAI) 33(01):4780–4789
Xie L, Yuille AL (2017) Genetic cnn. In: Proceedings of the IEEE international conference on computer vision (ICCV), 1379-1388
Zoph B, Vasudevan V, Shlens J, Le QV (2018) Learning transferable architectures for scalable image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 8697-8710
Iandola F, Moskewicz M, Karayev S, Girshick R, Darrell T, Keutzer K (2014) Densenet: implementing efficient convnet descriptor pyramids. In arXiv preprint arXiv:1404.1869
Han D, Kim J, Kim J (2017) Deep pyramidal residual networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 5927-5935
Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. In arXiv preprint arXiv:1704.04861
Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC (2018) Mobilenetv2: inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 4510-4520
Dong J, Cheng AC, Juan DC, Wei W, Sun M (2018) PPP-net: platform-aware progressive search for pareto-optimal neural architectures. In: International Conference on Learning Representations 2018 Workshop
Hsu C, Chang S, Liang J, Chou H, Liu C, Chang S, Pan J, Chen Y, Wei W, Juan DC (2018) Monas: multi-objective neural architecture search using reinforcement learning. In arXiv preprint arXiv:1806.10332
Zhang X, Jiang W, Shi Y, Hu J (2019) When Neural Architecture Search Meets Hardware Implementation: from hardware awareness to co-design. In: 2019 IEEE Computer Society Annual Symposium on VLSI (ISVLSI), 25-30
Kim S, Kwon H, Kwon E, Choi Y, Oh T, Kang S, MDARTS: Multi-objective Differentiable Neural Architecture Search. In: (2021) Design. Automation and Test in Europe Conference and Exhibition (DATE) 2021:1344–1349
Xue Y, Jiang P, Neri F, Liang J (2021) A multi-objective evolutionary approach based on graph-in-graph for neural architecture search of convolutional neural networks. Int J Neural Syst 31(09):2150035
Lyu B, Yuan H, Lu L, Zhang Y (2021) Resource-constrained neural architecture search on edge devices. IEEE Transactions Netw Sci Eng 9(1):134–142
Tan M, Chen B, Pang R, Vasudevan V, Sandler M, Howard A, Le QV (2019) Mnasnet: platform-aware neural architecture search for mobile. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 2820-2828
Wang Y, Ma Z, Wei X, Zheng S, Wang Y, Hong X (2022) ECCNAS: efficient crowd counting neural architecture search. ACM Transactions Multimedia Comput Commun Appl (TOMM) 18(1s):1–19
Tan H, Lai J, Liu Y, Song Y, Wang J, Chen M, Yan Y, Zhong L, Feng Q, Yang W (2022) Neural architecture search for real-time quality assessment of wearable multi-lead ECG on mobile devices. Biomed Signal Process Control 74:103495
Ding Z, Chen Y, Li N, Zhao D (2022) BNAS-v2: memory-efficient and performance-collapse-prevented broad neural architecture search. IEEE Trans Syst Man Cybern Syst 52(10):6259–6272
Ma A, Wan Y, Zhong Y, Wang J, Zhang L (2021) SceneNet: remote sensing scene classification deep learning network using multi-objective neural evolution architecture search. ISPRS J Photogramm Remote Sens 172:171–188
Kyriakides G, Margaritis K (2022) Evolving graph convolutional networks for neural architecture search. Neural Comput Appl 34:899–909
Zhang W, Fang J, Wang X, Liu W (2021) EfficientPose: efficient human pose estimation with neural architecture search. Comput Visual Media 7(3):335–347
Liu S, Zheng C, Lu K, Gao S, Wang N, Wang B, Zhang D, Zhang Xiao, Xu T (2021) Evsrnet: efficient video super-resolution with neural architecture search. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2480-2485
Stamoulis D, Ding R, Wang D, Lymberopoulos D, Priyantha B, Liu J, Marculescu D (2019) Single-path nas: designing hardware-efficient convnets in less than 4 hours. In European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD)
Liu C, Zoph B, Neumann M, Shlens J, Hua W, Li L, Li F, Yuille A, Huang J, Murphy K (2018) Progressive neural architecture search. In Progressive neural architecture search. In: Proceedings of the European conference on computer vision (ECCV), 19-34
White C, Zela A, Zela B, Liu Y, Hutter F (2021) How powerful are performance predictors in neural architecture search? Adv Neural Inform Process Syst 34:28454–28469
Kyriakides G, Margaritis K (2020) The effect of reduced training in neural architecture search. Neural Comput Appl 32:17321–17332
Yang Y, You S, Li H, Wang F, Qian C, Lin Z (2021) Towards improving the consistency, efficiency, and flexibility of differentiable neural architecture search. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 6667-6676
Ru B, Lyle C, Schut L, Fil M, Wilk M, Gal Y (2021) Speedy performance estimation for neural architecture search. Adv Neural Inform Process Syst 34
Wu B, Dai X, Zhang P, Wang Y, Sun F, Wu Y, Tian Y, Vajda P, Jia Y, Keutzer K (2019) Fbnet: hardware-aware efficient convnet design via differentiable neural architecture search. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 10734-10742
Chen Y, Yang T, Emer J, Sze V (2019) Eyeriss v2: a flexible accelerator for emerging deep neural networks on mobile devices. IEEE J Emerg Selected Topics Circuits Syst 9(2):292–308
Hegde K, Yu J, Agrawal R, Yan M, Pellauer M, Fletcher CW (2018) UCNN: exploiting computational reuse in deep neural networks via weight repetition, In 2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA), 674-687
Kwon H, Samajdar A, Krishna T (2018) MAERI: enabling flexible dataflow mapping over DNN accelerators via reconfigurable interconnects. ACM SIGPLAN Notices 53(2):461–475
Zhou Y, Ebrahimi S, Arık SÖ, Yu H, Liu H, Diamos G (2018) Resource-efficient neural architect. In arXiv preprint arXiv:1806.07912
Dong J, Cheng AC, Juan DC, Wei W, Sun M (2018) Dpp-net: device-aware progressive search for pareto-optimal neural architectures. In Proceedings of the European Conference on Computer Vision (ECCV), 517-531
Jiang W, Zhang X, Sha EH, Yang L, Zhuge Q, Shi Y, Hu J (2019) Accuracy vs. efficiency: achieving both through fpga-implementation aware neural architecture search. In Proceedings of the 56th Annual Design Automation Conference (DAC), (2019) 1-6
Cai L, Barneche AM, Herbout A, Foo CS, Lin J, Chandrasekhar VR, Sabry MM (2019) TEA-DNN: the quest for time-energy-accuracy co-optimized deep neural networks. In 2019 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED), 1-6
Benmeziane H, Maghraoui KE, Ouarnoughi H, Niar S, Wistuba M, Wang N (2021) a comprehensive survey on hardware-aware neural architecture search. In arXiv preprint arXiv:2101.09336
Cai H, Zhu L, Han S (2019) Proxylessnas: direct neural architecture search on target task and hardware. In: International Conference on Learning Representations (ICLR)
Liu H, Simonyan K, Yang Y (2019) Darts: differentiable architecture search. In: International Conference on Learning Representations (ICLR)
Ma Y, Cao Y, Vrudhula S, Seo J (2017) Optimizing loop operation and dataflow in fpga acceleration of deep convolutional neural networks. In: Proceedings of the 2017 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays (FPGA), 45-54
Weng J, Liu S, Wang Z, Dadu V, Nowatzki T (2020) A hybrid systolicdataflow architecture for inductive matrix algorithms. In 2020 IEEE International Symposium on High Performance Computer Architecture (HPCA), 703-716
Chollet F (2017) Xception: Deep learning with depthwise separable convolutions. In: In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 1251-1258
DeVries T, Taylor GW (2017) Improved regularization of convolutional neural networks with cutout. In arXiv preprint arXiv:1708.04552
Yamada Y, Iwamura M, Akiba T, Kise K (2019) Shakedrop regularization for deep residual learning. IEEE Access 7:186126–186136
Cai H, Gan C, Wang T, Zhang Z, Han S (2020) Once-for-all: train one network and specialize it for efficient deployment. In: International Conference on Learning Representations (ICLR)
Zhang X, Zhou X, Lin M, Sun J (2018) Shufflenet: an extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 6848-6856
Ma N, Zhang X, Zheng H, Sun J (2018) Shufflenet V2: practical guidelines for efficient CNN architecture design. In Proceedings of the European conference on computer vision (ECCV), 116-131
Tan M, Le QV (2019) EfficientNet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning (ICML), 6105-6114
Howard A, Sandler M, Chu G, Chen L, Chen B, Tan M, Wang W, Zhu Y, Pang R, Vasudevan V, Le QV, Adam H (2019) Searching for MobileNetV3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 1314-1324
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Yang, Z., Sun, Q. Toward efficient neural architecture search with dynamic mapping-adaptive sampling for resource-limited edge device. Neural Comput & Applic 35, 5553–5573 (2023). https://doi.org/10.1007/s00521-022-07984-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-07984-x