Abstract
Machine learning over distributed data collected by many clients has important applications in use cases where data privacy is a key concern or central data storage is not an option. Federated learning has introduced solutions for these scenarios, unlike the client-server approach, where all the training data is centralized in the server side, the clients, in a federated learning approach, perform machine learning updates locally over their data and the central server merely aggregates the resulting models without accessing the client’s local data. Article reviews the characteristics and learning objectives of the federated learning setting and gives an overview of the base algorithms that have been developed for FL, compare between FEDAVG and Consensus algorithm and shows the drawbacks and advantages of Consensus algorithm.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Savazzi, S., Nicoli, M., Bennis, M., Kianoush, S., Barbieri, L.: Opportunities of federated learning in connected, cooperative and automated industrial systems. IEEE Trans. Wireless Comm. 59(29) (2021)
Aledhari, M., Razzak, R., Parizi, R.M., Saeed, F.: Federated learning: a survey on enabling technologies, protocols, and applications. IEEE Access 8, 140699–140725 (2020)
Sery, T., Shlezinger, N., Cohen, K., Eldar, Y.C.: Over-the-air federated learning from heterogeneous data. IEEE Trans. Signal Process. 69 (2020)
Sayed, A.H.: Adaptation, learning, and optimization over networks. Found. Trends Mach. Learn. 7, 311–801 (2014)
Doku, R., Rawat, D.B., Liu, C.: Towards federated learning approach to determine data relevance in big data. In: IEEE 20th International Conference on Information Reuse and Integration for D ata Science (IRI), pp. 184–192, Los Angeles, USA (2019)
McMahan, H.B., Moore, E., Ramage, D., Hampson, S.: Communication-efficient learning of deep networks from decentralized data. arXiv preprint arXiv:1602.05629 (2016)
Liu, Y., et al.: FedCoin: a peer-to-peer payment system for federated learning. arXiv preprint arXiv:2002.11711 (2020)
Karimireddy, S.P., Kale, S., Mohri, M., Reddi, S., Stich, S., Suresh, A.T.: Scaffold: stochastic controlled averaging for federated learning. In: International Conference on Machine Learning, pp. 5132–5143. PMLR (2020)
Blot, M., Picard, D., Thome, N., Cord, M.: Distributed optimization for deep learning with gossip exchange. Neurocomputing 330, 287–296 (2019). https://doi.org/10.1016/j.neucom.2018.11.002
Savazzi, S., Nicoli, M., Rampa, V.: Federated learning with cooperating devices: a consensus approach for massive IoT networks. IEEE Internet Things J. (2020). https://doi.org/10.1109/JIOT.2020.2964162
Sayed, A.H., Tu, S.Y., Chen, J., Zhao, X.: Diffusion strategies for adaptation and learning over networks: an examination of distributed strategies and network behavior. IEEE Signal Proces. Mag. 30, 155–171 (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ali, A. (2023). Federated Learning Strategies Over Wireless Channels. In: Koucheryavy, Y., Aziz, A. (eds) Internet of Things, Smart Spaces, and Next Generation Networks and Systems. NEW2AN 2022. Lecture Notes in Computer Science, vol 13772. Springer, Cham. https://doi.org/10.1007/978-3-031-30258-9_47
Download citation
DOI: https://doi.org/10.1007/978-3-031-30258-9_47
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-30257-2
Online ISBN: 978-3-031-30258-9
eBook Packages: Computer ScienceComputer Science (R0)