skip to main content
10.1145/3318216.3363324acmconferencesArticle/Chapter ViewAbstractPublication PagessecConference Proceedingsconference-collections
poster

Distributed deep neural network training on edge devices

Published:07 November 2019Publication History

ABSTRACT

Deep Neural Network (Deep Learning) models have been traditionally trained on dedicated servers, after collecting data from various edge devices and sending them to the server. In recent years new methodologies have emerged for training models in a distributed manner over edge devices, keeping the data on the devices themselves. This allows for better data privacy and reduces the training costs. One of the main challenges for such methodologies is reducing the communication costs to and mainly from the edge devices. In this work we compare the two main methodologies used for distributed edge training: Federated Learning and Large Batch Training. For each of the methodologies we examine their convergence rates, communication costs, and final model performance. In addition, we present two techniques for compressing the communication between the edge devices, and examine their suitability for each one of the training methodologies.

References

  1. P. Goyal, P. Dollár, R. B. Girshick, P. Noordhuis, L. Wesolowski, A. Kyrola, A. Tulloch, Y. Jia, and K. He. Accurate, large minibatch SGD: training ImageNet in 1 hour. CoRR.Google ScholarGoogle Scholar
  2. N. S. Keskar, D. Mudigere, J. Nocedal, M. Smelyanskiy, and P. T. P. Tang. On large-batch training for deep learning: Generalization gap and sharp minima. ICLR, 2017.Google ScholarGoogle Scholar
  3. H. Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson Blaise Aguera y Arcas. Communication-Efficient Learning of Deep Networks from Decentralized Data.Google ScholarGoogle Scholar
  4. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep Residual Learning for Image Recognition.Google ScholarGoogle Scholar
  5. Jakub Konečný, H. Brendan McMahan, Felix X. Yu, Peter Richtárik, Ananda Theertha Suresh, Dave Bacon, "Federated Learning: Strategies for Improving Communication Efficiency"Google ScholarGoogle Scholar
  6. H. Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, Blaise Agüera y Arcas, "Communication-Efficient Learning of Deep Networks from Decentralized Data"Google ScholarGoogle Scholar
  7. Y. Lin, S. Han, H. Mao, Y. Wang, and W. J. Dally, "Deep gradient compression: Reducing the communication bandwidth for distributed training" arXiv preprint arXiv:1712.01887, 2017.Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    SEC '19: Proceedings of the 4th ACM/IEEE Symposium on Edge Computing
    November 2019
    455 pages
    ISBN:9781450367332
    DOI:10.1145/3318216

    Copyright © 2019 Owner/Author

    Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 7 November 2019

    Check for updates

    Qualifiers

    • poster

    Acceptance Rates

    SEC '19 Paper Acceptance Rate20of59submissions,34%Overall Acceptance Rate40of100submissions,40%

    Upcoming Conference

    SEC '24
    The Nineth ACM/IEEE Symposium on Edge Computing
    December 4 - 7, 2024
    Rome , Italy

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader