Abstract
The combination of federated learning and object detection techniques allows for the use of extensive datasets. This approach creates target detection models that are robust and adaptable. It also upholds strict privacy and security protocols. However, the computational resource variance among edge nodes poses significant constraints on the scalability of detection algorithms, thereby impinging on their accuracy and operational efficiency. This discrepancy compels clients, who might otherwise engage complex models, to resort to more simplified algorithms. To navigate this challenge, we introduce a federated learning framework enriched with a knowledge distillation strategy, specifically engineered for the object detection realm. This framework intelligently selects the largest model from the pool of client-contributed models and facilitates model upload and dissemination through a knowledge transfer mechanism orchestrated at the server side. Extensive experimental validations underscore the method’s proficiency in addressing the heterogeneity of detection algorithms within a varied environment, showcasing its viability for real-world application.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Ahn, S., Hu, S.X., Damianou, A., Lawrence, N.D., Dai, Z.: Variational information distillation for knowledge transfer. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9163–9171 (2019)
Chen, Y., Wang, C., Kim, B.: Federated learning with infrastructure resource limitations in vehicular object detection. In: 2021 IEEE/ACM Symposium on Edge Computing (SEC), pp. 366–370. IEEE (2021)
Everingham, M., Winn, J.: The pascal visual object classes challenge 2012 (voc2012) development kit. Pattern Anal. Stat. Model. Comput. Learn., Tech. Rep 2007(1-45), 5 (2012)
Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)
Jallepalli, D., Ravikumar, N.C., Badarinath, P.V., Uchil, S., Suresh, M.A.: Federated learning for object detection in autonomous vehicles. In: 2021 IEEE Seventh International Conference on Big Data Computing Service and Applications (BigDataService), pp. 107–114. IEEE (2021)
Jin, C., Chen, X., Gu, Y., Li, Q.: Feddyn: a dynamic and efficient federated distillation approach on recommender system. In: 2022 IEEE 28th International Conference on Parallel and Distributed Systems (ICPADS), pp. 786–793. IEEE (2023)
Kim, T., Lin, E., Lee, J., Lau, C., Mugunthan, V.: Navigating data heterogeneity in federated learning: a semi-supervised approach for object detection. Adv. Neural Inform. Process. Syst. 36 (2024)
Li, X., Jiang, M., Zhang, X., Kamp, M., Dou, Q.: Fedbn: Federated learning on non-iid features via local batch normalization. arXiv preprint arXiv:2102.07623 (2021)
Luo, J., Wu, X., Luo, Y., Huang, A., Huang, Y., Liu, Y., Yang, Q.: Real-world image datasets for federated learning. arXiv preprint arXiv:1910.11089 (2019)
McMahan, B., Moore, E., Ramage, D., Hampson, S., y Arcas, B.A.: Communication-efficient learning of deep networks from decentralized data. In: Artificial Intelligence and Statistics, pp. 1273–1282. PMLR (2017)
Park, W., Kim, D., Lu, Y., Cho, M.: Relational knowledge distillation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3967–3976 (2019)
Shu, C., Liu, Y., Gao, J., Yan, Z., Shen, C.: Channel-wise knowledge distillation for dense prediction. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 5311–5320 (2021)
Ultralytics: Yolov8 (2023). https://github.com/ultralytics/ultralytics
Wang, H., Li, Y., Xu, W., Li, R., Zhan, Y., Zeng, Z.: Dafkd: domain-aware federated knowledge distillation. In: Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, pp. 20412–20421 (2023)
Yang, L., Zhou, X., Li, X., Qiao, L., Li, Z., Yang, Z., Wang, G., Li, X.: Bridging cross-task protocol inconsistency for distillation in dense object detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 17175–17184 (2023)
Yang, Z., Li, Z., Gong, Y., Zhang, T., Lao, S., Yuan, C., Li, Y.: Rethinking knowledge distillation via cross-entropy. arXiv preprint arXiv:2208.10139 (2022)
Yim, J., Joo, D., Bae, J., Kim, J.: A gift from knowledge distillation: fast optimization, network minimization and transfer learning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4133–4141 (2017)
Zhang, J., Guo, S., Guo, J., Zeng, D., Zhou, J., Zomaya, A.: Towards data-independent knowledge transfer in model-heterogeneous federated learning. IEEE Trans. Comput. (2023)
Zhang, J., Zhou, J., Guo, J., Sun, X.: Visual object detection for privacy-preserving federated learning. IEEE Access 11, 33324–33335 (2023)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2025 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Hu, Y., Liu, G., Cheng, C. (2025). Federated Learning for Edge Heterogeneous Object Detection Algorithm. In: Cai, Z., Takabi, D., Guo, S., Zou, Y. (eds) Wireless Artificial Intelligent Computing Systems and Applications. WASA 2024. Lecture Notes in Computer Science, vol 14998. Springer, Cham. https://doi.org/10.1007/978-3-031-71467-2_7
Download citation
DOI: https://doi.org/10.1007/978-3-031-71467-2_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-71466-5
Online ISBN: 978-3-031-71467-2
eBook Packages: Computer ScienceComputer Science (R0)