skip to main content
10.1145/3220192.3220459acmconferencesArticle/Chapter ViewAbstractPublication PageshpdcConference Proceedingsconference-collections
research-article

Are Existing Knowledge Transfer Techniques Effective For Deep Learning on Edge Devices?

Published: 11 June 2018 Publication History

Abstract

With the emergence of edge computing paradigm, many applications such as image recognition and augmented reality require to perform machine learning (ML) and artificial intelligence (AI) tasks on edge devices. Most AI and ML models are large and computational-heavy, whereas edge devices are usually equipped with limited computational and storage resources. Such models can be compressed and reduced for deployment on edge devices, but they may loose their capability and not perform well. Recent works used knowledge transfer techniques to transfer information from a large network (termed teacher) to a small one (termed student) in order to improve the performance of the latter. This approach seems to be promising for learning on edge devices, but a thorough investigation on its effectiveness is lacking. This paper provides an extensive study on the performance (in both accuracy and convergence speed) of knowledge transfer, considering different student architectures and different techniques for transferring knowledge from teacher to student. The results show that the performance of KT does vary by architectures and transfer techniques. A good performance improvement is obtained by transferring knowledge from both the intermediate layers and last layer of the teacher to a shallower student. But other architectures and transfer techniques do not fare so well and some of them even lead to negative performance impact.

References

[1]
Martín Abadi. 2015. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. https://www.tensorflow.org/ Software available from tensorflow.org.
[2]
Jimmy Ba and Rich Caruana. 2014. Do deep nets really need to be deep?. In Advances in neural information processing systems. 2654--2662.
[3]
Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015).
[4]
Andrew G Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand, Marco Andreetto, and Hartwig Adam. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017).
[5]
Adriana Romero, Nicolas Ballas, Samira Ebrahimi Kahou, Antoine Chassang, Carlo Gatta, and Yoshua Bengio. 2014. Fitnets: Hints for thin deep nets. arXiv preprint arXiv:1412.6550 (2014).
[6]
Ragav Venkatesan and Baoxin Li. 2016. Diving deeper into mentee networks. arXiv preprint arXiv:1604.08220 (2016).

Cited By

View all
  • (2024)Using Cloning-GAN Architecture to Unlock the Secrets of Smart ManufacturingProcedia Computer Science10.1016/j.procs.2024.01.089232:C(890-902)Online publication date: 1-Jan-2024
  • (2024)Monitoring the growth of insect larvae using a regression convolutional neural network and knowledge transferEngineering Applications of Artificial Intelligence10.1016/j.engappai.2023.107358127(107358)Online publication date: Jan-2024
  • (2021)Machine Learning at the Network Edge: A SurveyACM Computing Surveys10.1145/346902954:8(1-37)Online publication date: 4-Oct-2021
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
HPDC '18: Proceedings of the 27th International Symposium on High-Performance Parallel and Distributed Computing
June 2018
25 pages
ISBN:9781450358996
DOI:10.1145/3220192
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

In-Cooperation

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 11 June 2018

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Deep learning
  2. Edge computing
  3. Knowledge transfer
  4. Neural networks

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

HPDC '18
Sponsor:

Acceptance Rates

Overall Acceptance Rate 61 of 390 submissions, 16%

Upcoming Conference

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)23
  • Downloads (Last 6 weeks)3
Reflects downloads up to 07 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Using Cloning-GAN Architecture to Unlock the Secrets of Smart ManufacturingProcedia Computer Science10.1016/j.procs.2024.01.089232:C(890-902)Online publication date: 1-Jan-2024
  • (2024)Monitoring the growth of insect larvae using a regression convolutional neural network and knowledge transferEngineering Applications of Artificial Intelligence10.1016/j.engappai.2023.107358127(107358)Online publication date: Jan-2024
  • (2021)Machine Learning at the Network Edge: A SurveyACM Computing Surveys10.1145/346902954:8(1-37)Online publication date: 4-Oct-2021
  • (2020)Convergence of Edge Computing and Deep Learning: A Comprehensive SurveyIEEE Communications Surveys & Tutorials10.1109/COMST.2020.297055022:2(869-904)Online publication date: Oct-2021
  • (2018)An Experimental Implementation of an Edge-based AI Engine with Edge-Cloud Coordination2018 18th International Symposium on Communications and Information Technologies (ISCIT)10.1109/ISCIT.2018.8587931(442-446)Online publication date: Sep-2018

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media