skip to main content
research-article

Personalized Federated Learning on Non-IID Data via Group-based Meta-learning

Published:22 March 2023Publication History
Skip Abstract Section

Abstract

Personalized federated learning (PFL) has emerged as a paradigm to provide a personalized model that can fit the local data distribution of each client. One natural choice for PFL is to leverage the fast adaptation capability of meta-learning, where it first obtains a single global model, and each client achieves a personalized model by fine-tuning the global one with its local data. However, existing meta-learning-based approaches implicitly assume that the data distribution among different clients is similar, which may not be applicable due to the property of data heterogeneity in federated learning. In this work, we propose a Group-based Federated Meta-Learning framework, called G-FML, which adaptively divides the clients into groups based on the similarity of their data distribution, and the personalized models are obtained with meta-learning within each group. In particular, we develop a simple yet effective grouping mechanism to adaptively partition the clients into multiple groups. Our mechanism ensures that each group is formed by the clients with similar data distribution such that the group-wise meta-model can achieve “personalization” at large. By doing so, our framework can be generalized to a highly heterogeneous environment. We evaluate the effectiveness of our proposed G-FML framework on three heterogeneous benchmarking datasets. The experimental results show that our framework improves the model accuracy by up to 13.15% relative to the state-of-the-art federated meta-learning.

REFERENCES

  1. [1] McMahan Brendan, Moore Eider, Ramage Daniel, Hampson Seth, and Arcas Blaise Agüera y. 2017. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics. 12731282.Google ScholarGoogle Scholar
  2. [2] Zhao Yue, Li Meng, Lai Liangzhen, Suda Naveen, Civin Damon, and Chandra Vikas. 2018. Federated learning with non-IID data. arXiv:1806.00582. Retrieved from https://arxiv.org/abs/1806.00582.Google ScholarGoogle Scholar
  3. [3] Li Xiang, Huang Kaixuan, Yang Wenhao, Wang Shusen, and Zhang Zhihua. 2020. On the convergence of FedAvg on non-IID data. In Proceedings of the 8th International Conference on Learning Representations.Google ScholarGoogle Scholar
  4. [4] Wu Chuhan, Wu Fangzhao, Cao Yang, Huang Yongfeng, and Xie Xing. 2021. FedGNN: Federated graph neural network for privacy-preserving recommendation. arXiv:2102.04925. Retrieved from https://arxiv.org/abs/2102.04925.Google ScholarGoogle Scholar
  5. [5] Smith Virginia, Chiang Chao-Kai, Sanjabi Maziar, and Talwalkar Ameet S.. 2017. Federated multi-task learning. In Proceedings of the Advances in Neural Information Processing Systems. 44244434.Google ScholarGoogle Scholar
  6. [6] Fallah Alireza, Mokhtari Aryan, and Ozdaglar Asuman E.. 2020. Personalized federated learning with theoretical guarantees: A model-agnostic meta-learning approach. In Proceedings of the Advances in Neural Information Processing Systems. 35573568.Google ScholarGoogle Scholar
  7. [7] Deng Yuyang, Kamani Mohammad Mahdi, and Mahdavi Mehrdad. 2020. Adaptive personalized federated learning. arXiv:2003.13461. Retrieved from https://arxiv.org/abs/2003.13461.Google ScholarGoogle Scholar
  8. [8] Dinh Canh T., Tran Nguyen H., and Nguyen Tuan Dung. 2020. Personalized federated learning with moreau envelopes. In Proceedings of the Advances in Neural Information Processing Systems. 2139421405.Google ScholarGoogle Scholar
  9. [9] Huang Yutao, Chu Lingyang, Zhou Zirui, Wang Lanjun, Liu Jiangchuan, Pei Jian, and Zhang Yong. 2021. Personalized cross-silo federated learning on non-IID data. In Proceedings of the AAAI Conference on Artificial Intelligence. 78657873.Google ScholarGoogle ScholarCross RefCross Ref
  10. [10] Chen Fei, Dong Zhenhua, Li Zhenguo, and He Xiuqiang. 2018. Federated meta-learning for recommendation. arXiv:1802.07876. Retrieved from https://arxiv.org/abs/1802.07876.Google ScholarGoogle Scholar
  11. [11] Jiang Yihan, Konečný Jakub, Rush Keith, and Kannan Sreeram. 2019. Improving federated learning personalization via model agnostic meta learning. arXiv:1909.12488. Retrieved from https://arxiv.org/abs/1909.12488.Google ScholarGoogle Scholar
  12. [12] Finn Chelsea, Abbeel Pieter, and Levine Sergey. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning. JMLR.org, 11261135.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. [13] Nichol Alex, Achiam Joshua, and Schulman John. 2018. On first-order meta-learning algorithms. arXiv:1803.02999. Retrieved from https://arxiv.org/abs/1803.02999.Google ScholarGoogle Scholar
  14. [14] Arivazhagan Manoj Ghuhan, Aggarwal Vinay, Singh Aaditya Kumar, and Choudhary Sunav. 2019. Federated learning with personalization layers. arXiv:1912.00818. Retrieved from https://arxiv.org/abs/1912.00818.Google ScholarGoogle Scholar
  15. [15] Shamsian Aviv, Navon Aviv, Fetaya Ethan, and Chechik Gal. 2021. Personalized federated learning using hypernetworks. In Proceedings of the 38th International Conference on Machine Learning. PMLR, 94899502.Google ScholarGoogle Scholar
  16. [16] Collins Liam, Hassani Hamed, Mokhtari Aryan, and Shakkottai Sanjay. 2021. Exploiting shared representations for personalized federated learning. In Proceedings of the 38th International Conference on Machine Learning. PMLR, 20892099.Google ScholarGoogle Scholar
  17. [17] Yang Qiang, Liu Yang, Chen Tianjian, and Tong Yongxin. 2019. Federated machine learning: Concept and applications. ACM Transactions on Intelligent Systems and Technology 10, 2 (2019), 119.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Li Tian, Sahu Anit Kumar, Talwalkar Ameet, and Smith Virginia. 2020. Federated learning: Challenges, methods, and future directions. IEEE Signal Processing Magazine 37, 3 (2020), 5060.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Kairouz Peter, McMahan H. Brendan, Avent Brendan, et al. 2021. Advances and open problems in federated learning. Foundations and Trends® in Machine Learning 14, 1–2 (2021), 1210.Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. [20] Mothukuri Viraaji, Parizi Reza M., Pouriyeh Seyedamin, Huang Yan, Dehghantanha Ali, and Srivastava Gautam. 2021. A survey on security and privacy of federated learning. Future Generation Computer Systems 115 (2021), 619640.Google ScholarGoogle ScholarCross RefCross Ref
  21. [21] Kulkarni Viraj, Kulkarni Milind, and Pant Aniruddha. 2020. Survey of personalization techniques for federated learning. In Proceedings of the 2020 4th World Conference on Smart Trends in Systems, Security, and Sustainability. IEEE, 794797.Google ScholarGoogle ScholarCross RefCross Ref
  22. [22] Tan Alysa Ziying, Yu Han, Cui Lizhen, and Yang Qiang. 2022. Towards personalized federated learning. IEEE Transactions on Neural Networks and Learning Systems (2022).Google ScholarGoogle ScholarCross RefCross Ref
  23. [23] Jeong Eunjeong, Oh Seungeun, Kim Hyesung, Park Jihong, Bennis Mehdi, and Kim Seong-Lyun. 2018. Communication-efficient on-device machine learning: Federated distillation and augmentation under non-IID private data. arXiv:1811.11479. Retrieved from https://arxiv.org/abs/1811.11479.Google ScholarGoogle Scholar
  24. [24] Duan Moming, Liu Duo, Chen Xianzhang, Liu Renping, Tan Yujuan, and Liang Liang. 2021. Self-balancing federated learning with global imbalanced data in mobile systems. IEEE Transactions on Parallel and Distributed Systems 32, 1 (2021), 5971.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. [25] Wang Hao, Kaplan Zakhary, Niu Di, and Li Baochun. 2020. Optimizing federated learning on non-IID data with reinforcement learning. In Proceedings of the IEEE INFOCOM. IEEE, 16981707.Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. [26] Yang Miao, Wang Ximin, Zhu Hongbin, Wang Haifeng, and Qian Hua. 2021. Federated learning with class imbalance reduction. In Proceedings of the 29th European Signal Processing Conference. IEEE, 21742178.Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Sattler Felix, Müller Klaus-Robert, and Samek Wojciech. 2021. Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints. IEEE Transactions on Neural Networks and Learning Systems 32, 8 (2021), 37103722.Google ScholarGoogle ScholarCross RefCross Ref
  28. [28] Ghosh Avishek, Chung Jichan, Yin Dong, and Ramchandran Kannan. 2020. An efficient framework for clustered federated learning. In Proceedings of the Advances in Neural Information Processing Systems. 1958619597.Google ScholarGoogle Scholar
  29. [29] Briggs Christopher, Fan Zhong, and Andras Peter. 2020. Federated learning with hierarchical clustering of local updates to improve training on non-IID data. In Proceedings of the 2020 International Joint Conference on Neural Networks. IEEE, 19.Google ScholarGoogle ScholarCross RefCross Ref
  30. [30] Duan Moming, Liu Duo, Ji Xinyuan, Liu Renping, Liang Liang, Chen Xianzhang, and Tan Yujuan. 2021. FedGroup: Efficient federated learning via decomposed similarity-based clustering. In Proceedings of the IEEE International Conference on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking.228237.Google ScholarGoogle Scholar
  31. [31] Li Daliang and Wang Junpu. 2019. FedMD: Heterogenous federated learning via model distillation. arXiv:1910.03581. Retrieved from https://arxiv.org/abs/1910.03581.Google ScholarGoogle Scholar
  32. [32] Chen Yiqiang, Qin Xin, Wang Jindong, Yu Chaohui, and Gao Wen. 2020. FedHealth: A federated transfer learning framework for wearable healthcare. IEEE Intelligent Systems 35, 4 (2020), 8393.Google ScholarGoogle ScholarCross RefCross Ref
  33. [33] Xie Ming, Long Guodong, Shen Tao, Zhou Tianyi, Wang Xianzhi, Jiang Jing, and Zhang Chengqi. 2021. Multi-center federated learning. arXiv:2108.08647. Retrieved from https://arxiv.org/abs/2108.08647.Google ScholarGoogle Scholar
  34. [34] Bui Duc, Malik Kshitiz, Goetz Jack, Liu Honglei, Moon Seungwhan, Kumar Anuj, and Shin Kang G.. 2019. Federated user representation learning. arXiv:1909.12535. Retrieved from https://arxiv.org/abs/1909.12535.Google ScholarGoogle Scholar
  35. [35] Liang Paul Pu, Liu Terrance, Liu Ziyin, Salakhutdinov Ruslan, and Morency Louis-Philippe. 2020. Think locally, act globally: Federated learning with local and global representations. arXiv:2001.01523. Retrieved from https://arxiv.org/abs/2001.01523.Google ScholarGoogle Scholar
  36. [36] Hospedales Timothy M., Antoniou Antreas, Micaelli Paul, and Storkey Amos J.. 2020. Meta-learning in neural networks: A survey. arXiv:2004.05439. Retrieved from https://arxiv.org/abs/2004.05439.Google ScholarGoogle Scholar
  37. [37] Lu Jiang, Gong Pinghua, Ye Jieping, and Zhang Changshui. 2020. Learning from very few samples: A survey. arXiv:2009.02653. Retrieved from https://arxiv.org/abs/2009.02653.Google ScholarGoogle Scholar
  38. [38] Lin Sen, Yang Guang, and Zhang Junshan. 2020. A collaborative learning framework via federated meta-learning. In Proceedings of the 40th IEEE International Conference on Distributed Computing Systems. IEEE, 289299.Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Khodak Mikhail, Balcan Maria-Florina, and Talwalkar Ameet S.. 2019. Adaptive gradient-based meta-learning methods. In Proceedings of the Advances in Neural Information Processing Systems. 59155926.Google ScholarGoogle Scholar
  40. [40] Vuorio Risto, Sun Shao-Hua, Hu Hexiang, and Lim Joseph J.. 2019. Multimodal model-agnostic meta-learning via task-aware modulation. In Proceedings of the Advances in Neural Information Processing Systems. 112.Google ScholarGoogle Scholar
  41. [41] Yao Huaxiu, Wei Ying, Huang Junzhou, and Li Zhenhui. 2019. Hierarchically structured meta-learning. In Proceedings of the 36th International Conference on Machine Learning. PMLR, 70457054.Google ScholarGoogle Scholar
  42. [42] Le Quoc V. and Mikolov Tomás. 2014. Distributed representations of sentences and documents. In Proceedings of the 31th International Conference on Machine Learning. JMLR.org, 11881196.Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Reimers Nils and Gurevych Iryna. 2019. Sentence-BERT: Sentence embeddings using siamese BERT-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing. Association for Computational Linguistics, 39803990.Google ScholarGoogle ScholarCross RefCross Ref
  44. [44] Caldas Sebastian, Wu Peter, Li Tian, Konečný Jakub, McMahan H. Brendan, Smith Virginia, and Talwalkar Ameet. 2018. LEAF: A benchmark for federated settings. arXiv:1812.01097. Retrieved from https://arxiv.org/abs/1812.01097.Google ScholarGoogle Scholar
  45. [45] Cohen Gregory, Afshar Saeed, Tapson Jonathan, and Schaik André van. 2017. EMNIST: Extending MNIST to handwritten letters. In Proceedings of the 2017 International Joint Conference on Neural Networks. IEEE, 29212926.Google ScholarGoogle ScholarCross RefCross Ref
  46. [46] Li Tian, Sahu Anit Kumar, Zaheer Manzil, Sanjabi Maziar, Talwalkar Ameet, and Smith Virginia. 2020. Federated optimization in heterogeneous networks. In Proceedings of the Machine Learning and Systems. mlsys.org.Google ScholarGoogle Scholar
  47. [47] Paszke Adam, Gross Sam, Massa Francisco, et al. 2019. PyTorch: An imperative style, high-performance deep learning library. In Proceedings of the Advances in Neural Information Processing Systems. 80248035.Google ScholarGoogle Scholar
  48. [48] Maaten Laurens Van der and Hinton Geoffrey. 2008. Visualizing data using t-SNE.Journal of Machine Learning Research 9, 11 (2008).Google ScholarGoogle Scholar

Index Terms

  1. Personalized Federated Learning on Non-IID Data via Group-based Meta-learning

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in

        Full Access

        • Published in

          cover image ACM Transactions on Knowledge Discovery from Data
          ACM Transactions on Knowledge Discovery from Data  Volume 17, Issue 4
          May 2023
          364 pages
          ISSN:1556-4681
          EISSN:1556-472X
          DOI:10.1145/3583065
          Issue’s Table of Contents

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 22 March 2023
          • Online AM: 23 August 2022
          • Accepted: 15 August 2022
          • Revised: 22 June 2022
          • Received: 3 November 2021
          Published in tkdd Volume 17, Issue 4

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • research-article

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader

        Full Text

        View this article in Full Text.

        View Full Text

        HTML Format

        View this article in HTML Format .

        View HTML Format