Abstract
Federated learning is a paradigm for distributed machine learning in which a central server interacts with a large number of remote devices to create the optimal global model. System and data heterogeneity are now the two largest impediments to federated learning. This work suggests a federated learning strategy based on stochastic gradient descent optimization as a solution to the problem of heterogeneity-induced slow convergence, or even non-convergence, of the global model. This work estimates the average global gradient using locally uploaded model parameters without computing the first derivative or updating global model parameters through gradient descent. Allowing the global model to be used with fewer communication rounds. Obtain faster and more reliable convergence results. In experiments simulating varying degrees of heterogeneous settings, the strategy proposed in this work delivered faster and more stable convergence than FedProx and FedAvg. This work offers a strategy that decreases the number of communication cycles on highly heterogeneous synthetic datasets by around 50% compared to FedProx, therefore considerably enhancing the stability and durability of federated learning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Hai, T., Zhou, J., Li, N., Jain, S.K., Agrawal, S., Dhaou, I.B.: Cloud-based bug tracking software defects analysis using deep learning. J. Cloud Comput. 11(1), 1–14 (2022)
Hai, T., Alsharif, S., Dhahad, H.A., Attia, E.A., Shamseldin, M.A., Ahmed, A.N.: The evolutionary artificial intelligence-based algorithm to find the minimum GHG emission via the integrated energy system using the MSW as fuel in a waste heat recovery plant. Sustain. Energy Technol. Assess. 53, 102531 (2022)
Mcmahan, B., Moore, E., Ramage, D., et al.: Communication-efficient learning of deep networks from decentralized data. In: Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, USA, pp. 1273−1282 (2017)
Wang, H., Yurochkin, M., Sun, Y., Papailiopoulos, D., Khazaeni, Y.: Federated learning with matched averaging. arXiv preprint arXiv:2002.06440 (2020)
Kopparapu, K., Lin, E., Zhao, J.: FedCD: improving performance in non-IID federated learning. arXiv preprint arXiv:2006.09637 (2020)
Yu, H., Yang, S., Zhu, S.: Parallel restarted SGD with faster convergence and less communication: demystifying why model averaging works for deep learning. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, pp. 5693–5700, July 2019
Liu, J., et al.: Adaptive asynchronous federated learning in resource-constrained edge computing. IEEE Trans. Mob. Comput. (2021)
Yu, H., Jin, R., Yang, S.: On the linear speedup analysis of communication efficient momentum SGD for distributed non-convex optimization. In: International Conference on Machine Learning, pp. 7184–7193. PMLR, May 2019
Guo, Y., Zhao, Z., He, K., Lai, S., Xia, J., Fan, L.: Efficient and flexible management for industrial internet of things: a federated learning approach. Comput. Netw. 192, 108122 (2021)
Wang, T., Liu, Y., Zheng, X., Dai, H.N., Jia, W., Xie, M.: Edge-based communication optimization for distributed federated learning. IEEE Trans. Netw. Sci. Eng. (2021)
Reddi, S., et al.: Adaptive federated optimization. arXiv preprint arXiv:2003.00295 (2020)
Yang, K., Fan, T., Chen, T., Shi, Y., Yang, Q.: A quasi-newton method based vertical federated learning framework for logistic regression. arXiv preprint arXiv:1912.00513 (2019)
Dhakal, S., Prakash, S., Yona, Y., Talwar, S., Himayat, N.: Coded federated learning. In: 2019 IEEE Globecom Workshops (GC Wkshps), pp. 1–6. IEEE, December 2019
Wang, C., Yang, Y., Zhou, P.: Towards efficient scheduling of federated mobile devices under computational and statistical heterogeneity. IEEE Trans. Parallel Distrib. Syst. 32(2), 394–410 (2020)
Malinovskiy, G., Kovalev, D., Gasanov, E., Condat, L., Richtarik, P.: From local SGD to local fixed-point methods for federated learning. In: International Conference on Machine Learning, pp. 6692–6701. PMLR, November 2020
Hanzely, F., Richtárik, P.: Federated learning of a mixture of global and local models. arXiv preprint arXiv:2002.05516 (2020)
Rothchild, D., et al.: FetchSGD: communication-efficient federated learning with sketching. In: International Conference on Machine Learning, pp. 8253–8265. PMLR, November 2020
Hai, T., Zhou, J., Muranaka, K.: An efficient fuzzy-logic based MPPT controller for grid-connected PV systems by farmland fertility optimization algorithm. Optik 169636 (2022)
Woodworth, B.E., Bullins, B., Shamir, O., Srebro, N.: The min-max complexity of distributed stochastic convex optimization with intermittent communication. In: Conference on Learning Theory, pp. 4386–4437. PMLR, July 2021
Tao, H., et al.: SDN-assisted technique for traffic control and information execution in vehicular adhoc networks. Comput. Electr. Eng. 108108 (2022)
Cui, Y., Jia, M., Lin, T.Y., Song, Y., Belongie, S.: Class-balanced loss based on effective number of samples. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9268–9277 (2019)
Hai, T., Said, N.M., Zain, J.M., Sajadi, S.M., Mahmoud, M.Z., Aybar, H.Ş.: ANN usefulness in building enhanced with PCM: Efficacy of PCM installation location. J. Build. Eng. 104914 (2022)
Li, T., Hu, S., Beirami, A., Smith, V.: DITTO: fair and robust federated learning through personalization. In: International Conference on Machine Learning, pp. 6357–6368. PMLR, July 2021
Li, T., Sahu, A.K., Zaheer, M., Sanjabi, M., Talwalkar, A., Smith, V.: Federated optimization in heterogeneous networks. Proc. Mach. Learn. Syst. 2, 429–450 (2020)
Charles, Z., Konečný, J.: Convergence and accuracy trade-offs in federated learning and meta-learning. In: International Conference on Artificial Intelligence and Statistics, pp. 2575–2583. PMLR, March 2021
Aono, Y., Hayashi, T., Wang, L., Moriai, S.: Privacy-preserving deep learning via additively homomorphic encryption. IEEE Trans. Inf. Forensics Secur. 13(5), 1333–1345 (2017)
Go, A., Bhayani, R., Huang, L.: Twitter sentiment classification using distant supervision. CS224N Project Rep. Stanford 1(12) (2009)
LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)
Cohen, G., Afshar, S., Tapson, J., Van Schaik, A.: EMNIST: extending MNIST to handwritten letters. In: 2017 International Joint Conference on Neural Networks (IJCNN), pp. 2921–2926. IEEE, May 2017
Tung, K.K., Tung, K.K.: Topics in Mathematical Modeling, vol. 10, p. 9781400884056. Princeton University Press, Princeton (2007)
Balles, L., Hennig, P.: Dissecting ADAM: the sign, magnitude and variance of stochastic gradients. In: International Conference on Machine Learning, pp. 404–413. PMLR, July 2018
Acknowledgement
This research was made possible with funding from the National Natural Science Foundation of China (No. 61862051), the Science and Technology Foundation of Guizhou Province (No. ZK[2022]549, No. [2019]1299), the Top-notch Talent Program of of Guizhou Province (No. KY[2018]080), the Natural Science Foundation of Education of Guizhou Province (No. [2019]203, KY[2019]067), and the Funds of Qiannan Normal University for Nationalities (No. qnsy2018003, No. qnsy2019rc09, No. qnsy2018JS013, No. qnsyrc201715).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhou, J., Zheng, M. (2023). Federated Learning with Class Balanced Loss Optimized by Implicit Stochastic Gradient Descent. In: Yusoff, M., Hai, T., Kassim, M., Mohamed, A., Kita, E. (eds) Soft Computing in Data Science. SCDS 2023. Communications in Computer and Information Science, vol 1771. Springer, Singapore. https://doi.org/10.1007/978-981-99-0405-1_9
Download citation
DOI: https://doi.org/10.1007/978-981-99-0405-1_9
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0404-4
Online ISBN: 978-981-99-0405-1
eBook Packages: Computer ScienceComputer Science (R0)