Skip to main content
Log in

A DNN inference acceleration algorithm combining model partition and task allocation in heterogeneous edge computing system

  • Published:
Peer-to-Peer Networking and Applications Aims and scope Submit manuscript

Abstract

Edge intelligence, as a new computing paradigm, aims to allocate Artificial Intelligence (AI)-based tasks partly on the edge to execute for reducing latency, consuming energy and improving privacy. As the most important technique of AI, Deep Neural Networks (DNNs) have been widely used in various fields. And for those DNN based tasks, a new computing scheme named DNN model partition can further reduce the execution time. This computing scheme partitions the DNN task into two parts, one will be executed on the end devices and the other will be executed on edge servers. However, in a complex edge computing system, it is difficult to coordinate DNN model partition and task allocation. In this work, we study this problem in the heterogeneous edge computing system. We first establish the mathematical model of adaptive DNN model partition and task offloading. The mathematical model contains a large number of binary variables, and the solution space will be too large to be solved directly in a multi-task scenario. Then we use dynamic programming and greedy strategy to reduce the solution space under the premise of a good solution, and propose our offline algorithm named GSPI. Then considering the actual situation, we subsequently proposed the online algorithm. Through our experiments and simulations, we proved that compared with end-only and server-only, our proposed GSPI algorithm can reduce the system time cost by 30% on average and the online algorithm can reduce the system time cost by 28% on average.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Parkhi OM, Vedaldi A, Zisserman A (2015) Deep face recognition. In BMVC

  2. Chen C, Seff A, Kornhauser AL, Xiao J (2015) Deepdriving: Learning affordance for direct perception in autonomous driving. 2015 IEEE International Conference on Computer Vision (ICCV) 2722–2730

  3. Chan W, Jaitly N, Le QV, Vinyals O (2016) Listen, attend and spell: A neural network for large vocabulary conversational speech recognition. 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 4960–4964

  4. Snyder T, Byrd G (2017) The internet of everything. Computer 50(6):8–9

    Article  Google Scholar 

  5. Pandey P, Singh S, Singh S (2010) Cloud computing. In ICWET

  6. Shi W, Cao J, Zhang Q, Li Y, Xu L (2016) Edge computing: Vision and challenges. IEEE Internet of Things J 3:637–646

    Article  Google Scholar 

  7. Long C, Cao Y, Jiang T, Zhang Q (2018) Edge computing framework for cooperative video processing in multimedia iot systems. IEEE Trans Multimedia 20:1126–1139

    Article  Google Scholar 

  8. Deschamps-Sonsino A (2018) Smarter homes. In Apress

  9. Alba E, Chicano F, Luque G (2016) Smart cities. In Lect Notes Comput Sci

  10. Zhou Z, Chen X, Li E, Zeng L, Luo K, Zhang J (2019) Edge intelligence: Paving the last mile of artificial intelligence with edge computing. Proc IEEE 107:1738–1762

    Article  Google Scholar 

  11. Kang Y, Hauswald J, Gao C, Rovinski A, Mudge TN, Mars J, Tang L (2017) Neurosurgeon: Collaborative intelligence between the cloud and mobile edge. In ASPLOS ’17

  12. Han S, Mao H, Dally WJ (2015) Deep compression: Compressing deep neural network with pruning, trained quantization and huffman coding. CoRR, abs/1510.00149

  13. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: Efficient convolutional neural networks for mobile vision applications. ArXiv, abs/1704.04861

  14. Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: Practical guidelines for efficient cnn architecture design. In ECCV

  15. Sandler M, Howard AG, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 4510–4520

  16. Zhang X, Zhou X, Lin M, Sun J (2018) Shufflenet: An extremely efficient convolutional neural network for mobile devices. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 6848–6856

  17. Xu M, Qian F, Zhu M, Huang F, Pushp S, Liu X (2020) Deepwear: Adaptive local offloading for on-wearable deep learning. IEEE Trans Mob Comput 19:314–330

    Article  Google Scholar 

  18. Li E, Zeng L, Zhou Z, Chen X (2020) Edge ai: On-demand accelerating deep neural network inference via edge computing. IEEE Trans Wirel Commun 19:447–457

    Article  Google Scholar 

  19. Teerapittayanon S, McDanel B, Kung HT (2016) Branchynet: Fast inference via early exiting from deep neural networks. 2016 23rd International Conference on Pattern Recognition (ICPR) 2464–2469

  20. Hu C, Bao WS, Wang D, Liu F (2019) Dynamic adaptive dnn surgery for inference acceleration on the edge. IEEE INFOCOM 2019 - IEEE Conference on Computer Communications 1423–1431

  21. Ko JH, Na T, Amir MF, Mukhopadhyay S (2018) Edge-host partitioning of deep neural networks with feature space encoding for resource-constrained internet-of-things platforms. 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS) 1–6

  22. Lin B, Huang Y, Zhang J, Hu J, Chen X, Li J (2020) Cost-driven off-loading for dnn-based applications over cloud, edge, and end devices. IEEE Trans Ind Inf 16:5456–5466

    Article  Google Scholar 

  23. Shi C, Chen L, Shen C, Song L, Xu J (2019) Privacy-aware edge computing based on adaptive dnn partitioning. 2019 IEEE Global Communications Conference (GLOBECOM) 1–6

  24. Mao Y, Zhang J, Letaief KB (2016) Dynamic computation offloading for mobile-edge computing with energy harvesting devices. IEEE J Sel Areas Commun 34:3590–3605

    Article  Google Scholar 

  25. Tran TX, Pompili D (2019) Joint task offloading and resource allocation for multi-server mobile-edge computing networks. IEEE Trans Veh Technol 68:856–868

    Article  Google Scholar 

  26. Mohammed T, Joe-Wong C, Babbar R, Francesco MD (2020) Distributed inference acceleration with adaptive dnn partitioning and offloading. IEEE INFOCOM 2020 - IEEE Conference on Computer Communications 854–863

  27. Huang Y, Wang F, Wang F, Liu J (2019) Deepar: A hybrid device-edge-cloud execution framework for mobile deep learning applications. IEEE INFOCOM 2019 - IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS) 892–897

  28. Qassim H, Feinzimer D, Verma A (2017) Residual squeeze vgg16. ArXiv, abs/1705.03004

  29. Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. Commun ACM 60:84–90

    Article  Google Scholar 

  30. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 770–778

  31. Szegedy C, Liu W, Jia Y, Sermanet P, Reed SE, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 1–9

Download references

Acknowledgements

This article was supported by the National Key Research And Development Plan(Grant No. 2018YFB2000505), National Natural Science Foundation of China (Grant No. 61806067) and Key Research and Development Project in Anhui Province(Grant No. 201904a06020024).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhigang Xu.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Shi, L., Xu, Z., Sun, Y. et al. A DNN inference acceleration algorithm combining model partition and task allocation in heterogeneous edge computing system. Peer-to-Peer Netw. Appl. 14, 4031–4045 (2021). https://doi.org/10.1007/s12083-021-01223-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12083-021-01223-1

Keywords

Navigation