skip to main content
10.1145/3589334.3645416acmconferencesArticle/Chapter ViewAbstractPublication PagesthewebconfConference Proceedingsconference-collections
research-article

FedDSE: Distribution-aware Sub-model Extraction for Federated Learning over Resource-constrained Devices

Published: 13 May 2024 Publication History

Abstract

Sub-model extraction based federated learning has emerged as a popular strategy for training models on resource-constrained devices. However, existing methods treat all clients equally and extract sub-models using predetermined rules, which disregard the statistical heterogeneity across clients and may lead to fierce competition among them. Specifically, this paper identifies that when making predictions, different clients tend to activate different neurons of the entire model related to their respective distributions. If highly activated neurons from some clients with one distribution are incorporated into the sub-model allocated to other clients with different distributions, they will be forced to fit the new distributions, which can hinder their activation over the previous clients and result in a performance reduction. Motivated by this finding, we propose a novel method called FedDSE, which can reduce the conflicts among clients by extracting sub-models based on the data distribution of each client. The core idea of FedDSE is to empower each client to adaptively extract neurons from the entire model based on their activation over the local dataset. We theoretically show that FedDSE can achieve an improved classification score and convergence over general neural networks with the ReLU activation function. Experimental results on various datasets and models show that FedDSE outperforms all state-of-the-art baselines.

Supplemental Material

MP4 File
Supplemental video

References

[1]
Maxwell Mbabilla Aladago and Lorenzo Torresani. [n. d.]. Slot Machines: Discovering Winning Combinations of Random Weights in Neural Networks. In38th International Conference on Machine Learning, ICML 2021.
[2]
Samiul Alam, Luyang Liu, Ming Yan, and Mi Zhang. [n. d.]. FedRolex: Model-Heterogeneous Federated Learning with Rolling Sub-Model Extraction. InAdvances in Neural Information Processing Systems 35 (NeurIPS 2022).
[3]
Sameer Bibikar, Haris Vikalo, Zhangyang Wang, and Xiaohan Chen. [n. d.]. Federated Dynamic Sparse Training: Computing Less, Communicating Less, Yet Learning Better. In36th AAAI Conference on Artificial Intelligence (AAAI 2022).
[4]
Han Cai, Chuang Gan, Ligeng Zhu, and Song Han. [n. d.]. TinyTL: Reduce Memory, Not Parameters for Efficient On-Device Learning. InNeurIPS 2020.
[5]
Sebastian Caldas and Jakub Kone?ný et al. [n. d.]. Expanding the Reach of Federated Learning by Reducing Client Resource Requirements. CoRR 2018 ([n. d.]).
[6]
Sebastian Caldas, Jakub Konecny, H Brendan McMahan, and Ameet Talwalkar. 2018. Expanding the reach of federated learning by reducing client resource requirements. arXiv preprint arXiv:1812.07210 (2018).
[7]
Timothy J. Castiglia, Anirban Das, Shiqiang Wang, and Stacy Patterson. [n. d.]. Compressed-VFL: Communication-Efficient Learning with Vertically Partitioned Data. InInternational Conference on Machine Learning (ICML 2022).
[8]
Enmao Diao, Jie Ding, and Vahid Tarokh. [n. d.]. HeteroFL: Computation and Communication Efficient Federated Learning for Heterogeneous Clients. In9th International Conference on Learning Representations (ICLR 2021).
[9]
Yucheng Ding, Chaoyue Niu, Fan Wu, Shaojie Tang, Chengfei Lyu, yanghe feng, and Guihai Chen. [n. d.]. Federated Submodel Optimization for Hot and Cold Data Features. InNeurIPS 2022.
[10]
Canh T. Dinh, Nguyen H. Tran, and Tuan Dung Nguyen. [n. d.]. Personalized Federated Learning with Moreau Envelopes. InNeurIPS 2020.
[11]
Cong Fang, Hangfeng He, Qi Long, and Weijie J. Su. [n. d.]. Exploring deep neural networks via layer-peeled model: Minority collapse in imbalanced training. Proceedings of the NationalAcademy of Sciences of the United States of America 2021 ([n. d.]).
[12]
Jonathan Frankle and Michael Carbin. [n. d.]. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks. InICLR 2019.
[13]
Graham Gobieski, Brandon Lucia, and Nathan Beckmann. [n. d.]. Intelligence Beyond the Edge: Inference on Intermittent Embedded Systems. InASPLOS 2019.
[14]
Song Han, Jeff Pool, John Tran, and William J. Dally. [n. d.]. Learning both Weights and Connections for Efficient Neural Network. InNeurIPS 2015.
[15]
Samuel Horváth, Stefanos Laskaridis, Mário Almeida, Ilias Leontiadis, Stylianos I. Venieris, and Nicholas D. Lane. [n. d.]. FjORD: Fair and Accurate Federated Learning under heterogeneous targets with Ordered Dropout. InNeurIPS 2021.
[16]
DONG-HYUN Hwang. [n. d.]. In-datacenter performance analysis of a tensor processing unit. ISCA 2017 ([n. d.]).
[17]
Berivan Isik, Francesco Pase, Deniz Gündüz, Tsachy Weissman, and Michele Zorzi. [n. d.]. Sparse Random Networks for Communication-Efficient Federated Learning. ICLR 2023 ([n. d.]).
[18]
Shaoxiong Ji, Wenqi Jiang, Anwar Walid, and Xue Li. 2022. Dynamic Sampling and Selective Masking for Communication-Efficient Federated Learning. IEEE Intell. Syst. 37, 2 (2022), 27--34.
[19]
Neel Kamal and Prasun Ghosal. 2018. Three Tier Architecture for IoT Driven Health Monitoring System Using Raspberry Pi. InIEEE International Symposium on Smart Electronic Systems.
[20]
Sai Praneeth Karimireddy, Satyen Kale, Mehryar Mohri, Sashank Reddi, Sebastian Stich, and Ananda Theertha Suresh. [n. d.]. SCAFFOLD: Stochastic controlled averaging for federated learning. InICML 2020.
[21]
Minjae Kim, Sangyoon Yu, Suhyun Kim, and Soo-Mook Moon. [n. d.]. DepthFL : Depthwise Federated Learning for Heterogeneous Clients. InICLR 2023.
[22]
Alex Krizhevsky, Geoffrey Hinton, et al. 2009. Learning multiple layers of features from tiny images. (2009).
[23]
Yann LeCun. 1998. The MNIST database of handwritten digits. http://yann. lecun. com/exdb/mnist/ (1998).
[24]
Seulki Lee, Bashima Islam, Yubo Luo, and Shahriar Nirjon. 2019. Intermittent Learning: On-Device Machine Learning on Intermittently Powered System. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 3, 4 (2019), 141:1--141:30.
[25]
Ang Li, Jingwei Sun, BinghuiWang, Lin Duan, Sicheng Li, Yiran Chen, and Hai Li. [n. d.]. LotteryFL: Personalized and Communication-Efficient Federated Learning with Lottery Ticket Hypothesis on Non-IID Datasets. CoRR 2020 ([n. d.]).
[26]
Ang Li, Jingwei Sun, Xiao Zeng, Mi Zhang, Hai Li, and Yiran Chen. [n. d.]. Fed- Mask: Joint Computation and Communication-Efficient Personalized Federated Learning via Heterogeneous Masking. InSenSys 2021.
[27]
Tian Li, Anit Kumar Sahu, and Manzil Zaheer et al. [n. d.]. Federated Optimization in Heterogeneous Networks. InMLSys 2020.
[28]
Tian Li, Maziar Sanjabi, Ahmad Beirami, and Virginia Smith. [n. d.]. Fair Resource Allocation in Federated Learning. InICLR 2020.
[29]
Dongping Liao, Xitong Gao, Yiren Zhao, and Chengzhong Xu. [n. d.]. Adaptive Channel Sparsity for Federated Learning under System Heterogeneity. InCVPR 2023.
[30]
Fangshuo Liao and Anastasios Kyrillidis. [n. d.]. On the Convergence of Shallow Neural Network Training with Randomly Masked Neurons. CoRR 2023 ([n. d.]).
[31]
Tao Lin, Lingjing Kong, Sebastian U. Stich, and Martin Jaggi. [n. d.]. Ensemble Distillation for Robust Model Fusion in Federated Learning. InNeurIPS 2020.
[32]
Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agüera y Arcas. 2017. Communication-Efficient Learning of Deep Networks from Decentralized Data. InAISTATS.
[33]
Mehryar Mohri, Gary Sivek, and Ananda Theertha Suresh. 2019. Agnostic federated learning. InInternational Conference on Machine Learning.
[34]
Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, et al. [n. d.]. Pytorch: An imperative style, high-performance deep learning library. NeurIPS 2019 ([n. d.]).
[35]
Ankit Pensia, Shashank Rajput, Alliot Nagle, Harit Vishwakarma, and Dimitris S. Papailiopoulos. [n. d.]. Optimal Lottery Tickets via Subset Sum: Logarithmic Over-Parameterization is Sufficient. InNeurIPS 2020.
[36]
Vivek Ramanujan, Mitchell Wortsman, Aniruddha Kembhavi, Ali Farhadi, and Mohammad Rastegari. [n. d.]. What's Hidden in a Randomly Weighted Neural Network?. InCVPR 2020.
[37]
Sejin Seo, Seung-Woo Ko, Jihong Park, et al. [n. d.]. Communication-Efficient and Personalized Federated Lottery Ticket Learning. InIEEE InternationalWorkshop on Signal Processing Advances in Wireless Communications 2021.
[38]
Muhammad Shafiq and Zhaoquan Gu. 2022. Deep Residual Learning for Image Recognition: A Survey. Applied Sciences 12, 18 (2022).
[39]
Yue Tan, Guodong Long, LU LIU, Tianyi Zhou, Qinghua Lu, Jing Jiang, and Chengqi Zhang. [n. d.]. FedProto: Federated Prototype Learning across Heterogeneous Clients. AAAI 2022 ([n. d.]).
[40]
Xueyang Tang, Song Guo, and Jingcai Guo. [n. d.]. Personalized Federated Learning with Contextualized Generalization. InIJCAI 2022.
[41]
Anish K. Vallapuram, Pengyuan Zhou, Young D. Kwon, Lik Hang Lee, Hengwei Xu, and Pan Hui. 2022. HideNseek: Federated Lottery Ticket via Server-side Pruning and Sign Supermask. CoRR abs/2206.04385 (2022).
[42]
Haozhao Wang, Yichen Li, Wenchao Xu, Ruixuan Li, Yufeng Zhan, and Zhigang Zeng. [n. d.]. DaFKD: Domain-aware Federated Knowledge Distillation. InCVPR 2023.
[43]
Jianyu Wang, Qinghua Liu, Hao Liang, Gauri Joshi, and H. Vincent Poor. [n. d.]. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization. InNeurIPS 2020.
[44]
An Xu, Wenqi Li, Pengfei Guo, Dong Yang, Holger Roth, Ali Hatamizadeh, Can Zhao, Daguang Xu, Heng Huang, and Ziyue Xu. [n. d.]. Closing the Generalization Gap of Cross-silo Federated Medical Image Segmentation. InCVPR 2022.
[45]
Hao Yu, Sen Yang, and Shenghuo Zhu. [n. d.]. Parallel Restarted SGD with Faster Convergence and Less Communication: Demystifying Why Model Averaging Works for Deep Learning. InAAAI 2019.
[46]
Hattie Zhou, Janice Lan, Rosanne Liu, and Jason Yosinski. [n. d.]. Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask. InNeurIPS 2019.
[47]
Zhuangdi Zhu, Junyuan Hong, and Jiayu Zhou. [n. d.]. Data-Free Knowledge Distillation for Heterogeneous Federated Learning. InICML 2021.

Cited By

View all
  • (2024)Is Aggregation the Only Choice? Federated Learning via Layer-wise Model RecombinationProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3637528.3671722(1096-1107)Online publication date: 25-Aug-2024

Index Terms

  1. FedDSE: Distribution-aware Sub-model Extraction for Federated Learning over Resource-constrained Devices

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    WWW '24: Proceedings of the ACM Web Conference 2024
    May 2024
    4826 pages
    ISBN:9798400701719
    DOI:10.1145/3589334
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 May 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. distribution-aware
    2. federated learning
    3. submodel extraction

    Qualifiers

    • Research-article

    Funding Sources

    • The research is supported under the National Key R\&D Program of China (2022ZD0160201) and the RIE2020 Industry Alignment Fund - Industry Collaboration Projects (IAF-ICP) Funding Initiative, as well as cash and in-kind contributions from the industry partner(s). This work is supported by National Natural Science Foundation of China under grants U1836204, U1936108, 62206102, and Science and Technology Support Program of Hubei Province under grant 2022BAA046 award number(s)

    Conference

    WWW '24
    Sponsor:
    WWW '24: The ACM Web Conference 2024
    May 13 - 17, 2024
    Singapore, Singapore

    Acceptance Rates

    Overall Acceptance Rate 1,899 of 8,196 submissions, 23%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)317
    • Downloads (Last 6 weeks)46
    Reflects downloads up to 01 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Is Aggregation the Only Choice? Federated Learning via Layer-wise Model RecombinationProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining10.1145/3637528.3671722(1096-1107)Online publication date: 25-Aug-2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media