Skip to main content

Federated Learning with Class Balanced Loss Optimized by Implicit Stochastic Gradient Descent

  • Conference paper
  • First Online:
Soft Computing in Data Science (SCDS 2023)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1771))

Included in the following conference series:

  • 275 Accesses

Abstract

Federated learning is a paradigm for distributed machine learning in which a central server interacts with a large number of remote devices to create the optimal global model. System and data heterogeneity are now the two largest impediments to federated learning. This work suggests a federated learning strategy based on stochastic gradient descent optimization as a solution to the problem of heterogeneity-induced slow convergence, or even non-convergence, of the global model. This work estimates the average global gradient using locally uploaded model parameters without computing the first derivative or updating global model parameters through gradient descent. Allowing the global model to be used with fewer communication rounds. Obtain faster and more reliable convergence results. In experiments simulating varying degrees of heterogeneous settings, the strategy proposed in this work delivered faster and more stable convergence than FedProx and FedAvg. This work offers a strategy that decreases the number of communication cycles on highly heterogeneous synthetic datasets by around 50% compared to FedProx, therefore considerably enhancing the stability and durability of federated learning.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Hai, T., Zhou, J., Li, N., Jain, S.K., Agrawal, S., Dhaou, I.B.: Cloud-based bug tracking software defects analysis using deep learning. J. Cloud Comput. 11(1), 1–14 (2022)

    Article  Google Scholar 

  2. Hai, T., Alsharif, S., Dhahad, H.A., Attia, E.A., Shamseldin, M.A., Ahmed, A.N.: The evolutionary artificial intelligence-based algorithm to find the minimum GHG emission via the integrated energy system using the MSW as fuel in a waste heat recovery plant. Sustain. Energy Technol. Assess. 53, 102531 (2022)

    Google Scholar 

  3. Mcmahan, B., Moore, E., Ramage, D., et al.: Communication-efficient learning of deep networks from decentralized data. In: Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, USA, pp. 1273−1282 (2017)

    Google Scholar 

  4. Wang, H., Yurochkin, M., Sun, Y., Papailiopoulos, D., Khazaeni, Y.: Federated learning with matched averaging. arXiv preprint arXiv:2002.06440 (2020)

    Google Scholar 

  5. Kopparapu, K., Lin, E., Zhao, J.: FedCD: improving performance in non-IID federated learning. arXiv preprint arXiv:2006.09637 (2020)

    Google Scholar 

  6. Yu, H., Yang, S., Zhu, S.: Parallel restarted SGD with faster convergence and less communication: demystifying why model averaging works for deep learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, pp. 5693–5700, July 2019

    Google Scholar 

  7. Liu, J., et al.: Adaptive asynchronous federated learning in resource-constrained edge computing. IEEE Trans. Mob. Comput. (2021)

    Google Scholar 

  8. Yu, H., Jin, R., Yang, S.: On the linear speedup analysis of communication efficient momentum SGD for distributed non-convex optimization. In: International Conference on Machine Learning, pp. 7184–7193. PMLR, May 2019

    Google Scholar 

  9. Guo, Y., Zhao, Z., He, K., Lai, S., Xia, J., Fan, L.: Efficient and flexible management for industrial internet of things: a federated learning approach. Comput. Netw. 192, 108122 (2021)

    Article  Google Scholar 

  10. Wang, T., Liu, Y., Zheng, X., Dai, H.N., Jia, W., Xie, M.: Edge-based communication optimization for distributed federated learning. IEEE Trans. Netw. Sci. Eng. (2021)

    Google Scholar 

  11. Reddi, S., et al.: Adaptive federated optimization. arXiv preprint arXiv:2003.00295 (2020)

    Google Scholar 

  12. Yang, K., Fan, T., Chen, T., Shi, Y., Yang, Q.: A quasi-newton method based vertical federated learning framework for logistic regression. arXiv preprint arXiv:1912.00513 (2019)

    Google Scholar 

  13. Dhakal, S., Prakash, S., Yona, Y., Talwar, S., Himayat, N.: Coded federated learning. In: 2019 IEEE Globecom Workshops (GC Wkshps), pp. 1–6. IEEE, December 2019

    Google Scholar 

  14. Wang, C., Yang, Y., Zhou, P.: Towards efficient scheduling of federated mobile devices under computational and statistical heterogeneity. IEEE Trans. Parallel Distrib. Syst. 32(2), 394–410 (2020)

    Article  Google Scholar 

  15. Malinovskiy, G., Kovalev, D., Gasanov, E., Condat, L., Richtarik, P.: From local SGD to local fixed-point methods for federated learning. In: International Conference on Machine Learning, pp. 6692–6701. PMLR, November 2020

    Google Scholar 

  16. Hanzely, F., Richtárik, P.: Federated learning of a mixture of global and local models. arXiv preprint arXiv:2002.05516 (2020)

    Google Scholar 

  17. Rothchild, D., et al.: FetchSGD: communication-efficient federated learning with sketching. In: International Conference on Machine Learning, pp. 8253–8265. PMLR, November 2020

    Google Scholar 

  18. Hai, T., Zhou, J., Muranaka, K.: An efficient fuzzy-logic based MPPT controller for grid-connected PV systems by farmland fertility optimization algorithm. Optik 169636 (2022)

    Google Scholar 

  19. Woodworth, B.E., Bullins, B., Shamir, O., Srebro, N.: The min-max complexity of distributed stochastic convex optimization with intermittent communication. In: Conference on Learning Theory, pp. 4386–4437. PMLR, July 2021

    Google Scholar 

  20. Tao, H., et al.: SDN-assisted technique for traffic control and information execution in vehicular adhoc networks. Comput. Electr. Eng. 108108 (2022)

    Google Scholar 

  21. Cui, Y., Jia, M., Lin, T.Y., Song, Y., Belongie, S.: Class-balanced loss based on effective number of samples. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9268–9277 (2019)

    Google Scholar 

  22. Hai, T., Said, N.M., Zain, J.M., Sajadi, S.M., Mahmoud, M.Z., Aybar, H.Ş.: ANN usefulness in building enhanced with PCM: Efficacy of PCM installation location. J. Build. Eng. 104914 (2022)

    Google Scholar 

  23. Li, T., Hu, S., Beirami, A., Smith, V.: DITTO: fair and robust federated learning through personalization. In: International Conference on Machine Learning, pp. 6357–6368. PMLR, July 2021

    Google Scholar 

  24. Li, T., Sahu, A.K., Zaheer, M., Sanjabi, M., Talwalkar, A., Smith, V.: Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2, 429–450 (2020)

    Google Scholar 

  25. Charles, Z., Konečný, J.: Convergence and accuracy trade-offs in federated learning and meta-learning. In: International Conference on Artificial Intelligence and Statistics, pp. 2575–2583. PMLR, March 2021

    Google Scholar 

  26. Aono, Y., Hayashi, T., Wang, L., Moriai, S.: Privacy-preserving deep learning via additively homomorphic encryption. IEEE Trans. Inf. Forensics Secur. 13(5), 1333–1345 (2017)

    Google Scholar 

  27. Go, A., Bhayani, R., Huang, L.: Twitter sentiment classification using distant supervision. CS224N Project Rep. Stanford 1(12) (2009)

    Google Scholar 

  28. LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)

    Article  Google Scholar 

  29. Cohen, G., Afshar, S., Tapson, J., Van Schaik, A.: EMNIST: extending MNIST to handwritten letters. In: 2017 International Joint Conference on Neural Networks (IJCNN), pp. 2921–2926. IEEE, May 2017

    Google Scholar 

  30. Tung, K.K., Tung, K.K.: Topics in Mathematical Modeling, vol. 10, p. 9781400884056. Princeton University Press, Princeton (2007)

    Google Scholar 

  31. Balles, L., Hennig, P.: Dissecting ADAM: the sign, magnitude and variance of stochastic gradients. In: International Conference on Machine Learning, pp. 404–413. PMLR, July 2018

    Google Scholar 

Download references

Acknowledgement

This research was made possible with funding from the National Natural Science Foundation of China (No. 61862051), the Science and Technology Foundation of Guizhou Province (No. ZK[2022]549, No. [2019]1299), the Top-notch Talent Program of of Guizhou Province (No. KY[2018]080), the Natural Science Foundation of Education of Guizhou Province (No. [2019]203, KY[2019]067), and the Funds of Qiannan Normal University for Nationalities (No. qnsy2018003, No. qnsy2019rc09, No. qnsy2018JS013, No. qnsyrc201715).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jincheng Zhou .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhou, J., Zheng, M. (2023). Federated Learning with Class Balanced Loss Optimized by Implicit Stochastic Gradient Descent. In: Yusoff, M., Hai, T., Kassim, M., Mohamed, A., Kita, E. (eds) Soft Computing in Data Science. SCDS 2023. Communications in Computer and Information Science, vol 1771. Springer, Singapore. https://doi.org/10.1007/978-981-99-0405-1_9

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-0405-1_9

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-0404-4

  • Online ISBN: 978-981-99-0405-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics