skip to main content
10.1145/3447548.3467309acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article

Fed2: Feature-Aligned Federated Learning

Authors Info & Claims
Published:14 August 2021Publication History

ABSTRACT

Federated learning learns from scattered data by fusing collaborative models from local nodes. However, conventional coordinate-based model averaging by FedAvg ignored the random information encoded per parameter and may suffer from structural feature misalignment. In this work, we propose Fed2, a feature-aligned federated learning framework to resolve this issue by establishing a firm structure-feature alignment across the collaborative models. Fed2 is composed of two major designs: First, we design a feature-oriented model structure adaptation method to ensure explicit feature allocation in different neural network structures. Applying the structure adaptation to collaborative models, matchable structures with similar feature information can be initialized at the very early training stage. During the federated learning process, we then propose a feature paired averaging scheme to guarantee aligned feature distribution and maintain no feature fusion conflicts under either IID or non-IID scenarios. Eventually, Fed2 could effectively enhance the federated learning convergence performance under extensive homo- and heterogeneous settings, providing excellent convergence speed, accuracy, and computation/communication efficiency.

Skip Supplemental Material Section

Supplemental Material

KDD21-rst2039.mp4

mp4

105.3 MB

References

  1. A. Gonzalez-Garcia, D. Modolo, and V. Ferrari. Do semantic parts emerge in convolutional neural networks? International Journal of Computer Vision, 126(5):476--494, 2018.Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Andrew G Howard and et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv:1704.04861, 2017.Google ScholarGoogle Scholar
  3. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097--1105, 2012.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Tian Li, Anit Kumar Sahu, Ameet Talwalkar, and Virginia Smith. Federated learning: Challenges, methods, and future directions. IEEE Signal Processing Magazine, 37(3):50--60, 2020.Google ScholarGoogle ScholarCross RefCross Ref
  5. T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith. Federated optimization in heterogeneous networks. In Proceedings of the 3rd MLSys Conference, 2018.Google ScholarGoogle Scholar
  6. X. Li, K. Huang, W. Yang, S. Wang, and Z. Zhang. On the convergence of fedavg on non-iid data. In Proceedings of the 8th International Conference on Learning Representations (ICLR), 2019.Google ScholarGoogle Scholar
  7. H Brendan McMahan and et al. Communication-efficient learning of deep networks from decentralized data. International Conference on Artificial Intelligence and Statistics (AISTATS), 2017.Google ScholarGoogle Scholar
  8. H. Mostafa. Robust federated learning through representation matching and adaptive hyper-parameters. arXiv:1912.13075, 2019.Google ScholarGoogle Scholar
  9. Z. Qin, F. Yu, C. Liu, and X. Chen. How convolutional neural network see the world-a survey of convolutional neural network visualization methods. arXiv preprint:1804.11191, 2018.Google ScholarGoogle Scholar
  10. Z. Qin, F. Yu, C. Liu, and X. Chen. Functionality-oriented convolutional filter pruning. In Proceedings of 30th British Machine Vision Conference (BMVC), 2019.Google ScholarGoogle Scholar
  11. W. Samek, A. Binder, G. Montavon, S. Lapuschkin, and K. Müller. Evaluating the visualization of what a deep neural network has learned. IEEE transactions on neural networks and learning systems, 28(11):2660--2673, 2016.Google ScholarGoogle Scholar
  12. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556, 2014.Google ScholarGoogle Scholar
  13. S. P. Singh and M. Jaggi. Model fusion via optimal transport. arXiv:1910.05653, 2019.Google ScholarGoogle Scholar
  14. H. Wang and et al. Federated learning with matched averaging. In Proceedings of the 8th International Conference on Learning Representations (ICLR), 2020.Google ScholarGoogle Scholar
  15. Yuxin Wu and Kaiming He. Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV), 2018.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Qiang Yang and et al. Federated learning. Synthesis Lectures on Artificial Intelligence and Machine Learning, 13(3):1--207, 2019.Google ScholarGoogle ScholarCross RefCross Ref
  17. Qiang Yang, Yang Liu, Tianjian Chen, and Yongxin Tong. Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology (TIST), 10(2):1--19, 2019.Google ScholarGoogle Scholar
  18. Jason Yosinski, Jeff Clune, Anh Nguyen, Thomas Fuchs, and Hod Lipson. Understanding neural networks through deep visualization. arXiv:1506.06579, 2015.Google ScholarGoogle Scholar
  19. Fuxun Yu, Zhuwei Qin, and Xiang Chen. Distilling critical paths in convolutional neural networks. arXiv:1811.02643, 2018.Google ScholarGoogle Scholar
  20. M. Yurochkin and et al. Probabilistic federated neural matching. 2018.Google ScholarGoogle Scholar
  21. M. Yurochkin and et al. Bayesian nonparametric federated learning of neural networks. In Proceedings of the 36th International Conference on Machine Learning (ICML), pages 7252--7261, 2019.Google ScholarGoogle Scholar
  22. Xiangyu Zhang and et al. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018.Google ScholarGoogle Scholar
  23. Yue Zhao and et al. Federated learning with non-iid data. arXiv:1806.00582, 2018.Google ScholarGoogle Scholar
  24. B. Zhou and et al. Learning deep features for discriminative localization. In Proceedings of Computer Vision and Pattern Recognition (CVPR), pages 2921--2929, 2016.Google ScholarGoogle Scholar

Index Terms

  1. Fed2: Feature-Aligned Federated Learning

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      KDD '21: Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining
      August 2021
      4259 pages
      ISBN:9781450383325
      DOI:10.1145/3447548

      Copyright © 2021 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 14 August 2021

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate1,133of8,635submissions,13%

      Upcoming Conference

      KDD '24

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader