skip to main content
10.1145/3512527.3531372acmconferencesArticle/Chapter ViewAbstractPublication PagesicmrConference Proceedingsconference-collections
research-article

FedNKD: A Dependable Federated Learning Using Fine-tuned Random Noise and Knowledge Distillation

Published: 27 June 2022 Publication History

Abstract

Multimedia retrieval models need the ability to extract useful information from large-scale data for clients. As an important part of multimedia retrieval, image classification model directly affects the efficiency and effect of multimedia retrieval. We need a lot of data to train a image classification model applied to multimedia retrieval task. However, with the protection of data privacy, the data used to train the model often needs to be kept on the client side. Federated learning is proposed to use data from all clients to train one model while protecting privacy. When federated learning is applied, the distribution of data across different clients varies greatly. Disregarding this problem yields a final model with unstable performance. To enable federated learning to work dependably in the real world with complex data environments, we propose FedNKD, which utilizes knowledge distillation and random noise. The superior knowledge of each client is distilled into a central server to mitigate the instablity caused by Non-IID data. Importantly, a synthetic dataset is created by some random noise through back propagation of neural networks. The synthetic dataset will contain the abstract features of the real data. Then we will use this synthetic dataset to realize the knowledge distillation while protecting users' privacy. In our experimental scenarios, FedNKD outperforms existing representative algorithms by about 1.5% in accuracy.

Supplementary Material

MP4 File (ICMR22-fp103.mp4)
We intend to improve the performance of federated learning algorithms in the case of highly unbalanced data distribution. For example, each client has only a few categories of data. We use knowledge distillation to distillate the expertise of different clients into the model of the central server, so that it has a good performance in the global data. To protect client data privacy, we use noise data and fine-tune it so that knowledge distillation can be performed without compromising user privacy. Finally, our FedNKD method achieved the best results in some scenarios.

References

[1]
Michael Adams, Suriadi Suriadi, Akhil Kumar, and Arthur H. M. ter Hofstede. [n.d.]. Flexible Integration of Blockchain with Business Process Automation: A Federated Architecture. In Advanced Information Systems Engineering - CAiSE Forum 2020, Grenoble, France, June 8--12, 2020, Proceedings. 1--13. https://doi.org/10.1007/978--3-030--58135-0_1
[2]
Asra Aslam. [n.d.]. Object Detection for Unseen Domains while Reducing Response Time using Knowledge Transfer in Multimedia Event Processing. In Proceedings of the 2020 on International Conference on Multimedia Retrieval, ICMR 2020, Dublin, Ireland, June 8--11, 2020. 373--377. https://doi.org/10.1145/3372278.3391936
[3]
Kallista A. Bonawitz, Hubert Eichner, Wolfgang Grieskamp, Dzmitry Huba, Alex Ingerman, Vladimir Ivanov, Chloé Kiddon, Jakub Konecný, Stefano Mazzocchi, Brendan McMahan, Timon Van Overveldt, David Petrou, Daniel Ramage, and Jason Roselander. [n.d.]. Towards Federated Learning at Scale: System Design. In Proceedings of Machine Learning and Systems 2019, MLSys 2019, Stanford, CA, USA, March 31 - April 2, 2019.
[4]
Petra Budíková, Jan Sedmidubský, and Pavel Zezula. [n.d.]. Efficient Indexing of 3D Human Motions. In ICMR '21: International Conference on Multimedia Retrieval, Taipei, Taiwan, August 21--24, 2021. 10--18. https://doi.org/10.1145/3460426.3463646
[5]
Petr Byvshev, Pascal Mettes, and Yu Xiao. [n.d.]. Heterogeneous Non-Local Fusion for Multimodal Activity Recognition. In Proceedings of the 2020 on International Conference on Multimedia Retrieval, ICMR 2020, Dublin, Ireland, June 8--11, 2020. 63--72. https://doi.org/10.1145/3372278.3390675
[6]
Fabio Carrara, Giuseppe Amato, Fabrizio Falchi, and Claudio Gennaro. [n.d.]. Continuous ODE-defined Image Features for Adaptive Retrieval. In Proceedings of the 2020 on International Conference on Multimedia Retrieval, ICMR 2020, Dublin, Ireland, June 8--11, 2020. 198--206. https://doi.org/10.1145/3372278.3390690
[7]
Walter Priesnitz Filho, Carlos Ribeiro, and Thomas Zefferer. [n.d.]. Towards Privacy-preserving Attribute Aggregation in Federated eID Systems. In Proceedings of the CAiSE'16 Forum, at the 28th International Conference on Advanced Information Systems Engineering (CAiSE 2016), Ljubljana, Slovenia, June 13--17, 2016. 97--104.
[8]
Ian J. Goodfellow, Jonathon Shlens, and Christian Szegedy. [n.d.]. Explaining and Harnessing Adversarial Examples. In 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7--9, 2015, Conference Track Proceedings.
[9]
Qiushan Guo, Xinjiang Wang, Yichao Wu, Zhipeng Yu, Ding Liang, Xiaolin Hu, and Ping Luo. [n.d.]. Online Knowledge Distillation via Collaborative Learning. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, June 13--19, 2020. 11017--11026. https://doi.org/10.1109/CVPR42600.2020.01103
[10]
Chaoyang He, Murali Annavaram, and Salman Avestimehr. [n.d.]. Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6--12, 2020, virtual.
[11]
Geoffrey E. Hinton, Oriol Vinyals, and Jeffrey Dean. 2015. Distilling the Knowledge in a Neural Network. (2015). http://arxiv.org/abs/1503.02531
[12]
Peter Kairouz, H. Brendan McMahan, Brendan Avent, Aurélien Bellet, Mehdi Bennis, Arjun Nitin Bhagoji, Kallista A. Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, Rafael G. L. D'Oliveira, Hubert Eichner, Salim El Rouayheb, David Evans, Josh Gardner, Zachary Garrett, Adrià Gascón, Badih Ghazi, Phillip B. Gibbons, Marco Gruteser, Zaïd Harchaoui, Chaoyang He, Lie He, Zhouyuan Huo, Ben Hutchinson, Justin Hsu, Martin Jaggi, Tara Javidi, Gauri Joshi, Mikhail Khodak, Jakub Konecný, Aleksandra Korolova, Farinaz Koushanfar, Sanmi Koyejo, Tancrède Lepoint, Yang Liu, Prateek Mittal, Mehryar Mohri, Richard Nock, Ayfer Özgür, Rasmus Pagh, Hang Qi, Daniel Ramage, Ramesh Raskar, and etc. Mariana Raykova. [n.d.]. Advances and Open Problems in Federated Learning. Found. Trends Mach. Learn. ([n. d.]), 1--210. https://doi.org/10.1561/2200000083
[13]
Sai Praneeth Karimireddy, Satyen Kale, Mehryar Mohri, Sashank J. Reddi, Sebastian U. Stich, and Ananda Theertha Suresh. [n.d.]. SCAFFOLD: Stochastic Controlled Averaging for Federated Learning. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13--18 July 2020, Virtual Event. 5132--5143.
[14]
Jakub Konecný, H. Brendan McMahan, Felix X. Yu, Peter Richtárik, Ananda Theertha Suresh, and Dave Bacon. 2016. Federated Learning: Strategies for Improving Communication Efficiency. (2016). http://arxiv.org/abs/1610.05492
[15]
Alex Krizhevsky, Ilya Sutskever, and Geoffrey E. Hinton. 2012. ImageNet Classification with Deep Convolutional Neural Networks. In NIPS. 1106--1114.
[16]
Tian Li, Anit Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith. [n.d.]. Federated Optimization in Heterogeneous Networks. In Proceedings of Machine Learning and Systems 2020, MLSys 2020, Austin, TX, USA, March 2--4, 2020.
[17]
Xiang Li, Kaixuan Huang, Wenhao Yang, Shusen Wang, and Zhihua Zhang. 2020. On the Convergence of FedAvg on Non-IID Data. In 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26--30, 2020.
[18]
Wei Yang Bryan Lim, Nguyen Cong Luong, Dinh Thai Hoang, Yutao Jiao, Ying-Chang Liang, Qiang Yang, Dusit Niyato, and Chunyan Miao. [n.d.]. Federated Learning in Mobile Edge Networks: A Comprehensive Survey. IEEE Commun. Surv. Tutorials ([n. d.]), 2031--2063. https://doi.org/10.1109/COMST.2020.2986024
[19]
Tao Lin, Lingjing Kong, Sebastian U. Stich, and Martin Jaggi. [n.d.]. Ensemble Distillation for Robust Model Fusion in Federated Learning. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6--12, 2020, virtual.
[20]
Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Agüera y Arcas. [n.d.]. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, AISTATS 2017, 20--22 April 2017, Fort Lauderdale, FL, USA. 1273--1282.
[21]
Seyed-Iman Mirzadeh, Mehrdad Farajtabar, Ang Li, Nir Levine, Akihiro Matsukawa, and Hassan Ghasemzadeh. [n.d.]. Improved Knowledge Distillation via Teacher Assistant. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7--12, 2020. 5191--5198.
[22]
Viraaji Mothukuri, Reza M. Parizi, Seyedamin Pouriyeh, Yan Huang, Ali Dehghantanha, and Gautam Srivastava. [n.d.]. A survey on security and privacy of federated learning. Future Gener. Comput. Syst. ([n. d.]), 619--640. https://doi.org/10.1016/j.future.2020.10.007
[23]
Milad Nasr, Reza Shokri, and Amir Houmansadr. [n.d.]. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning. In 2019 IEEE Symposium on Security and Privacy, SP 2019, San Francisco, CA, USA, May 19--23, 2019. 739--753. https://doi.org/10.1109/SP.2019.00065
[24]
Yu-Shu Ni, Chia-Chi Tsai, Jiun-In Guo, Jenq-Neng Hwang, Bo-Xun Wu, Po-Chi Hu, Ted T. Kuo, Po-Yu Chen, and Hsien-Kai Kuo. [n.d.]. Summary of the 2021 Embedded Deep Learning Object Detection Model Compression Competition for Traffic in Asian Countries. In ICMR '21: International Conference on Multimedia Retrieval, Taipei, Taiwan, August 21--24, 2021. 244--249. https://doi.org/10.1145/3460426.3466932
[25]
Wonpyo Park, Dongju Kim, Yan Lu, and Minsu Cho. [n.d.]. Relational Knowledge Distillation. In IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16--20, 2019. 3967--3976. https://doi.org/10.1109/CVPR.2019.00409
[26]
Felix Sattler, Simon Wiedemann, Klaus-Robert Müller, and Wojciech Samek. 2020. Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data. IEEE Trans. Neural Networks Learn. Syst. (2020), 3400--3413. https://doi.org/10.1109/TNNLS.2019.2944481
[27]
Joachim Scherer and Gerd Kiparski. 2018. Buchbesprechungen. Feiler, Lukas / Forgó, Nikolaus / Weigl, Michaela: The Eu General Data Protection Regulation (Gdpr): A Commentary. Comput. und Recht (2018), 69--70. https://doi.org/10.9785/cr-2018--340626
[28]
Jiawei Su, Danilo Vasconcellos Vargas, and Kouichi Sakurai. 2019. One Pixel Attack for Fooling Deep Neural Networks. IEEE Trans. Evol. Comput. (2019), 828--841. https://doi.org/10.1109/TEVC.2019.2890858
[29]
Laurens Van der Maaten and Geoffrey Hinton. 2008. Visualizing data using t-SNE. Journal of machine learning research (2008).
[30]
Jianyu Wang, Qinghua Liu, Hao Liang, Gauri Joshi, and H. Vincent Poor. [n.d.]. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6--12, 2020, virtual.
[31]
Shiqiang Wang, Tiffany Tuor, Theodoros Salonidis, Kin K. Leung, Christian Makaya, Ting He, and Kevin Chan. [n.d.]. Adaptive Federated Learning in Resource Constrained Edge Computing Systems. IEEE J. Sel. Areas Commun. ([n. d.]), 1205--1221. https://doi.org/10.1109/JSAC.2019.2904348
[32]
Lingxi Xie, Richang Hong, Bo Zhang, and Qi Tian. [n.d.]. Image Classification and Retrieval are ONE. In Proceedings of the 5th ACM on International Conference on Multimedia Retrieval, Shanghai, China, June 23--26, 2015. 3--10. https://doi.org/10.1145/2671188.2749289
[33]
Guodong Xu, Ziwei Liu, Xiaoxiao Li, and Chen Change Loy. [n.d.]. Knowledge Distillation Meets Self-supervision. In Computer Vision - ECCV 2020 - 16th European Conference, Glasgow, UK, August 23--28, 2020, Proceedings, Part IX. 588--604. https://doi.org/10.1007/978--3-030--58545--7_34
[34]
Qiang Yang, Yang Liu, Tianjian Chen, and Yongxin Tong. [n.d.]. Federated Machine Learning: Concept and Applications. ACM Trans. Intell. Syst. Technol. ([n. d.]), 12:1--12:19. https://doi.org/10.1145/3298981
[35]
Yue Zhao, Meng Li, Liangzhen Lai, Naveen Suda, Damon Civin, and Vikas Chandra. 2018. Federated Learning with Non-IID Data. (2018). http://arxiv.org/abs/1806.00582
[36]
Zhuangdi Zhu, Junyuan Hong, and Jiayu Zhou. [n.d.]. Data-Free Knowledge Distillation for Heterogeneous Federated Learning. In Proceedings of the 38th International Conference on Machine Learning, ICML 2021, 18--24 July 2021, Virtual Event. 12878--12889.

Cited By

View all
  • (2024)Communication Efficiency and Non-Independent and Identically Distributed Data Challenge in Federated Learning: A Systematic Mapping StudyApplied Sciences10.3390/app1407272014:7(2720)Online publication date: 24-Mar-2024
  • (2024)A comprehensive survey of federated transfer learning: challenges, methods and applicationsFrontiers of Computer Science10.1007/s11704-024-40065-x18:6Online publication date: 23-Jul-2024
  • (2023)FedCE: Personalized Federated Learning Method based on Clustering EnsemblesProceedings of the 31st ACM International Conference on Multimedia10.1145/3581783.3612217(1625-1633)Online publication date: 26-Oct-2023

Index Terms

  1. FedNKD: A Dependable Federated Learning Using Fine-tuned Random Noise and Knowledge Distillation

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      ICMR '22: Proceedings of the 2022 International Conference on Multimedia Retrieval
      June 2022
      714 pages
      ISBN:9781450392389
      DOI:10.1145/3512527
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 27 June 2022

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. Non-IID
      2. federated learning
      3. knowledge distillation

      Qualifiers

      • Research-article

      Funding Sources

      Conference

      ICMR '22
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 254 of 830 submissions, 31%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)41
      • Downloads (Last 6 weeks)10
      Reflects downloads up to 28 Feb 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Communication Efficiency and Non-Independent and Identically Distributed Data Challenge in Federated Learning: A Systematic Mapping StudyApplied Sciences10.3390/app1407272014:7(2720)Online publication date: 24-Mar-2024
      • (2024)A comprehensive survey of federated transfer learning: challenges, methods and applicationsFrontiers of Computer Science10.1007/s11704-024-40065-x18:6Online publication date: 23-Jul-2024
      • (2023)FedCE: Personalized Federated Learning Method based on Clustering EnsemblesProceedings of the 31st ACM International Conference on Multimedia10.1145/3581783.3612217(1625-1633)Online publication date: 26-Oct-2023

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media