Abstract
Efficient and flexible cloud computing is widely used in distributed systems. However, in the Internet of Things (IoT) environment with heterogeneous capabilities, the performance of cloud computing may decline due to limited communication resources. As located closer to the end, edge computing is used to replace cloud computing to provide timely and stable services. To accomplish distributed system and privacy preserving, Federated Learning (FL) has been combined with edge computing. However, due to the large number of clients, the amount of data transmitted will also grow exponentially. How to reduce the communication overhead in FL is still a big problem. As a major method to reduce the communication overhead, compressing the transmission parameters can effectively reduce the communication overhead. However, the existing methods do not consider the possible internal relationship between neurons. In this paper, we propose Neuron Pruning-Based FL for communication-efficient distributed training, which is a model pruning method to compress model parameters transmitted in FL. In contrast to the previous methods, we use dimensionality reduction method as the importance factor of neurons, and take advantage of the correlation between them to carry out model pruning. Our analysis results show that NPBFL can reduce communication overhead while maintaining classification accuracy.
This research was supported by the National Key R &D Program of China under Grant No. 2022YFB3102304 and in part by National Natural Science Foundation of China Grants (6222510562001057).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Peng, C., Hu, Q., Chen, J., Kang, K., Li, F., Zou, X.: Energy-efficient device selection in federated edge learning. In: 30th International Conference on Computer Communications and Networks, ICCCN 2021, Athens, Greece, 19–22 July 2021, pp. 1–9. IEEE (2021). https://doi.org/10.1109/ICCCN52240.2021.9522303
Yang, D., Cheng, Z., Zhang, W., Zhang, H., Shen, X.: Burst-aware time-triggered flow scheduling with enhanced multi-CQF in time-sensitive networks. IEEE/ACM Trans. Networking 31(6), 2809–2824 (2023). https://doi.org/10.1109/TNET.2023.3264583
Tange, K., Donno, M.D., Fafoutis, X., Dragoni, N.: A systematic survey of industrial Internet of Things security: requirements and fog computing opportunities. IEEE Commun. Surv. Tutorials 22(4), 2489–2520 (2020). https://doi.org/10.1109/COMST.2020.3011208
Al-Garadi, M.A., Mohamed, A., Al-Ali, A.K., Du, X., Ali, I., Guizani, M.: A survey of machine and deep learning methods for Internet of Things (IoT) security. IEEE Commun. Surv. Tutorials 22(3), 1646–1685 (2020). https://doi.org/10.1109/COMST.2020.2988293
Sha, J., Basara, N., Freedman, J., Xu, H.: FLOR: a federated learning-based music recommendation engine. In: 31st International Conference on Computer Communications and Networks, ICCCN 2022, Honolulu, HI, USA, 25–28 July 2022, pp. 1–2. IEEE (2022). https://doi.org/10.1109/ICCCN54977.2022.9868921
Zhang, W., et al.: Optimizing federated learning in distributed industrial IoT: a multi-agent approach. IEEE J. Sel. Areas Commun. 39(12), 3688–3703 (2021). https://doi.org/10.1109/JSAC.2021.3118352
Yang, D., et al.: DetFed: dynamic resource scheduling for deterministic federated learning over time-sensitive networks. IEEE Trans. Mobile Comput. (2023). Early Access, https://doi.org/10.1109/TMC.2023.3303017
Mo, Z., Gao, Z., Zhao, C., Lin, Y.: FedDQ: a communication-efficient federated learning approach for internet of vehicles. J. Syst. Archit. 131, 102690 (2022). https://doi.org/10.1016/j.sysarc.2022.102690
Peng, K., Zhang, H., Zhao, B., Liu, P.: Edge-cloud collaborative computation offloading for federated learning in smart city. In: IEEE International Conference on Dependable, Autonomic and Secure Computing, International Conference on Pervasive Intelligence and Computing, International Conference on Cloud and Big Data Computing, International Conference on Cyber Science and Technology Congress, DASC/PiCom/CBDCom/CyberSciTech 2022, Falerna, Italy, 12–15 September 2022, pp. 1–7. IEEE (2022). https://doi.org/10.1109/DASC/PiCom/CBDCom/Cy55231.2022.9927848
Otoum, Y., Wan, Y., Nayak, A.: Federated transfer learning-based IDS for the internet of medical things (IoMT). In: IEEE Globecom 2021 Workshops, Madrid, Spain, 7–11 December 2021, pp. 1–6. IEEE (2021). https://doi.org/10.1109/GCWkshps52748.2021.9682118
Deng, Y., et al.: SHARE: shaping data distribution at edge for communication-efficient hierarchical federated learning. In: 41st IEEE International Conference on Distributed Computing Systems, ICDCS 2021, Washington DC, USA, 7–10 July 2021, pp. 24–34. IEEE (2021). https://doi.org/10.1109/ICDCS51616.2021.00012
Luo, S., Fan, P., Xing, H., Luo, L., Yu, H.: Eliminating communication bottlenecks in cross-device federated learning with in-network processing at the edge. In: IEEE International Conference on Communications, ICC 2022, Seoul, Korea, 16–20 May 2022, pp. 4601–4606. IEEE (2022). https://doi.org/10.1109/ICC45855.2022.9838381
Liu, L., Zhang, J., Song, S., Letaief, K.B.: Communication-efficient federated distillation with active data sampling. In: IEEE International Conference on Communications, ICC 2022, Seoul, Korea, 16–20 May 2022, pp. 201–206. IEEE (2022). https://doi.org/10.1109/ICC45855.2022.9839214
Qu, L., Song, S., Tsui, C.: FedDQ: communication-efficient federated learning with descending quantization. In: IEEE Global Communications Conference, GLOBECOM 2022, Rio de Janeiro, Brazil, 4–8 December 2022, pp. 281–286. IEEE (2022). https://doi.org/10.1109/GLOBECOM48099.2022.10001205
Lian, Z., Wang, W., Su, C.: COFEL: communication-efficient and optimized federated learning with local differential privacy. In: ICC 2021 - IEEE International Conference on Communications, Montreal, QC, Canada, 14–23 June 2021, pp. 1–6. IEEE (2021). https://doi.org/10.1109/ICC42927.2021.9500632
Vahidian, S., Morafah, M., Lin, B.: Personalized federated learning by structured and unstructured pruning under data heterogeneity. In: 41st IEEE International Conference on Distributed Computing Systems Workshops, ICDCSW 2021, Washington, DC, USA, 7–10 July 2021, pp. 27–34. IEEE (2021). https://doi.org/10.1109/ICDCSW53096.2021.00012
Magnússon, S., Ghadikolaei, H.S., Li, N.: On maintaining linear convergence of distributed learning and optimization under limited communication. IEEE Trans. Signal Process. 68, 6101–6116 (2020). https://doi.org/10.1109/TSP.2020.3031073
Li, Y., Gu, S., Mayer, C., Gool, L.V., Timofte, R.: Group sparsity: the hinge between filter pruning and decomposition for network compression. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020, pp. 8015–8024. Computer Vision Foundation/IEEE (2020). https://doi.org/10.1109/CVPR42600.2020.00804, https://openaccess.thecvf.com/content_CVPR_2020/html/Li_Group_Sparsity_The_Hinge_Between_Filter_Pruning_and_Decomposition_for_CVPR_2020_paper.html
Zhao, K., Zhang, X., Han, Q., Cheng, M.: Dependency aware filter pruning. CoRR abs/2005.02634 (2020). https://arxiv.org/abs/2005.02634
Louizos, C., Welling, M., Kingma, D.P.: Learning sparse neural networks through l\({}_{{0}}\) regularization. CoRR abs/1712.01312 (2017). http://arxiv.org/abs/1712.01312
Lin, M., et al.: HRank: filter pruning using high-rank feature map. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020, pp. 1526–1535. Computer Vision Foundation/IEEE (2020). https://doi.org/10.1109/CVPR42600.2020.00160, https://openaccess.thecvf.com/content_CVPR_2020/html/Lin_HRank_Filter_Pruning_Using_High-Rank_Feature_Map_CVPR_2020_paper.html
Chen, Y., Dai, X., Liu, M., Chen, D., Yuan, L., Liu, Z.: Dynamic convolution: attention over convolution kernels. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020
Wang, W., Zhu, M., Zeng, X., Ye, X., Sheng, Y.: Malware traffic classification using convolutional neural network for representation learning. In: 2017 International Conference on Information Networking (ICOIN), pp. 712–717 (2017). https://doi.org/10.1109/ICOIN.2017.7899588
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Guan, J., Wang, P., Yao, S., Zhang, J. (2024). Neuron Pruning-Based Federated Learning for Communication-Efficient Distributed Training. In: Tari, Z., Li, K., Wu, H. (eds) Algorithms and Architectures for Parallel Processing. ICA3PP 2023. Lecture Notes in Computer Science, vol 14490. Springer, Singapore. https://doi.org/10.1007/978-981-97-0859-8_4
Download citation
DOI: https://doi.org/10.1007/978-981-97-0859-8_4
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-97-0858-1
Online ISBN: 978-981-97-0859-8
eBook Packages: Computer ScienceComputer Science (R0)