Elsevier

Computer Networks

Volume 183, 24 December 2020, 107569
Computer Networks

Customized Federated Learning for accelerated edge computing with heterogeneous task targets

https://doi.org/10.1016/j.comnet.2020.107569Get rights and content

Abstract

As a dominant edge intelligence technique, Federated Learning (FL) can reduce the data transmission volume, shorten the communication latency and improve the collaboration efficiency among end-devices and edge servers. Existing works on FL-based edge computing only take device- and resource-heterogeneity into consideration under a fixed loss-minimization objective. As heterogeneous end-devices are usually assigned with various tasks with different target accuracies, task heterogeneity is also a significant issue and has not yet been investigated. To this end, we propose a Customized FL (CuFL) algorithm with an adaptive learning rate to tailor for heterogeneous accuracy requirements and to accelerate the local training process. We also present a fair global aggregation strategy for the edge server to minimize the variance of accuracy gaps among heterogeneous end-devices. We rigorously analyze the convergence property of the CuFL algorithm in theory. We also verify the feasibility and effectiveness of the CuFL algorithm in the vehicle classification task. Evaluation results demonstrate that our algorithm performs better in terms of the accuracy rate, training time, and fairness during aggregation than existing efforts.

Introduction

With the rapid development of smart applications (e.g., smart transportation, smart healthcare, and smart agriculture), wireless devices at the network termination are assigned with various intelligent tasks (e.g., identification, prediction, and policy optimization). As most scenarios are user-interactive and latency-sensitive, real-time response is of great importance to the user experience. In the Mobile Edge Computing (MEC) paradigm, communication latencies of end-devices can be reduced since MEC servers (i.e., edge servers) are configured with certain computing capabilities and are deployed in the vicinity of end-devices [1], [2], [3]. Considering that resource-constrained end-devices have difficulties in tackling complex computing tasks, it has become an inevitable trend to coordinate multiple end-devices to execute edge tasks intelligently [4]. In traditional edge intelligence methods, end-devices need to offload local data to the adjacent MEC server, which naturally costs transmission time and poses threats to data privacy [5].

In order to further reduce the communication latency and protect data privacy, Google initiates a collaborative machine learning framework, termed Federated Learning (FL) [6], [7]. Without exposing any raw data, distributed end-devices can independently train a specific machine learning model (e.g., support vector machine) based on their local data sets. After several iterations of local training, end-devices will upload model parameters to the MEC server, which aggregates diverse model parameters and sends the updated parameters back after global aggregation [8]. Compared with conventional edge intelligence approaches, FL makes some breakthroughs: (i) FL breaks the isolation of individual data and facilitates the joint execution of dispersed end-devices; (ii) it naturally avoids data leakage and protects privacy of end users; (iii) FL can reduce the training time and save energy due to the fact that only model parameters are transmitted instead of the whole data set [9]. Due to the incomparable advantages, FL has aroused widespread attention recently.

Existing FL methods over mobile edge networks mainly focus on the heterogeneities in terms of devices (e.g., CPU-cycle frequency) and resources (e.g., power, wireless channels) [10]. To tackle the above heterogeneity bottlenecks, a number of works have been proposed to maximize the training accuracy of the predefined model while reducing training time and energy consumption [11].

Moreover, existing FL-based edge computing methods generally assume that different end-devices share a consistent training target. Considering that heterogeneous end-devices may be assigned with various tasks and different targets (e.g., different levels of training accuracies) [12], it is necessary to investigate the individualized FL scheme to adapt to various training accuracies. Under the individualized FL framework, there is no need to cater for an extravagant accuracy. Instead, the early satisfaction of the predefined accuracy requirement can make end-devices timely quit the training process, and evidently reduce training latency and save energy.

In order to accelerate the FL process while adapting to heterogeneous tasks with different levels of accuracy requirements, we propose a Customized FL (CuFL) algorithm with an adaptive learning rate. In our algorithm, each end-device has a distinctive accuracy target. We utilize a large learning rate to quickly search for the target accuracy, and utilize a small learning rate when the current accuracy approaches the target one. Once training targets are satisfied, end-devices can quit training processes before predefined deadlines. Besides the optimization in the local training stage, we design a parameter aggregation method that minimizes the variance between current accuracies and target accuracies to maintain the fair global aggregation.

Our main contributions are summarized as follows:

  • Considering that heterogeneous devices are usually assigned with various tasks under individualized accuracy targets, we propose the CuFL algorithm to accelerate the FL process while ensuring that all end-devices can satisfy their specific task requirements. To our knowledge, CuFL is the first algorithm that can tackle the multi-dimensional heterogeneities on devices, resources and tasks simultaneously.

  • In order to accelerate the local training process, we innovatively introduce an adaptive learning rate to tailor for end-devices’ individualized training requirements. By defining a gap function between the current accuracy and the target accuracy, the local learning rate is set to a large value initially to speed up the model convergence and gradually decreases as the gap becomes smaller.

  • To further accelerate the local model training for end-devices, we propose an early termination scheme to shorten the training time by cutting down the aggregation rounds. In the early termination scheme, end-devices can quit the FL process in advance when they satisfy the accuracy requirements. As a result, the energy cost is reduced and communication resources will be abundant for the remaining devices.

  • At the side of the MEC server, we optimize the global aggregation method. In order to achieve a fair parameter aggregation at the MEC server, we introduce a fairness coefficient to minimize the variance between the current and target accuracies.

  • We rigorously analyze the convergence property of the CuFL algorithm in theory. We also verify the effectiveness of CuFL on the vehicle classification task. Evaluation results show its superiorities in terms of the accuracy rate, training time and fairness during aggregation.

The reminder of this paper is organized as follows. Section 2 reviews the existing researches about FL. Section 3 presents the system model and problem formulation. Section 4 presents the CuFL algorithm and analyzes its convergence property. In Section 5, we conduct experiments to evaluate the CuFL algorithm. Finally, Section 6 concludes this paper.

Section snippets

Related work

FL emerges as a novel technology that aggregates distributed end-devices to collaboratively learn an intelligent algorithm with the aid of a central MEC server [13]. As FL includes end-devices’ local training and the MEC server’s global aggregation, we thus review related works in these two aspects.

Preliminaries on FL

We consider a wireless edge network that consists of one MEC server and multiple end-devices. Let K denote the set of end-devices, where K=|K| denotes the device number. Each end-device k(kK) owns its local data set Dk, the size of which is denoted as Dk. The total data size of all participating devices is D=k=1KDk.

As shown in Fig. 1, FL includes the local training stage and the global aggregation stage [8]. During the local training stage, smart devices train a specific machine learning

CuFL algorithm

Based on the aforementioned optimizations, the redesigned FL architecture is shown in Fig. 3. In the CuFL algorithm, the device-edge system aims at minimizing the completion time in executing FL. Except for model parameters, end-devices transmit extra data sizes and loss bounds to the MEC server. During the global aggregation, we replace the original FedAvg method with our fair aggregation method. The details of the CuFL algorithm are described in Section 4.1.

Performance evaluation

In this section, we conduct three sets of simulations to evaluate the algorithm performances in terms of the model accuracy, training time, and fairness during aggregation. We select the classic FedAvg algorithm [7] and the fair-aggregation q-FFL method [24] as the baselines for comparison purposes.

Conclusion

In this paper, we proposed the CuFL algorithm with an adaptive learning rate to minimize the total learning time. Considering the task heterogeneity of end-devices, CuFL allows end-devices to quit the training in advance once they have met their distinctive accuracy requirements. We also introduced a fairness coefficient in the global aggregation step to minimize the variance between the current and target accuracies. We proved the deterministic convergence property theoretically. Extensive

CRediT authorship contribution statement

Hui Jiang: Conceptualization, Methodology, Writing - original draft, Investigation. Min Liu: Conceptualization, Project administration. Bo Yang: Writing - review & editing. Qingxiang Liu: Software, Validation. Jizhong Li: Project administration. Xiaobing Guo: Supervision.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgment

This work was supported by the National Natural Science Foundation of China under Grant No. 61732017, No. 62072436, No. 61872028.

Hui Jiang received her B.S. degree from the Department of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing, China, in 2017. Since 2017, she is currently a Ph.D. candidate at the Networking Technology Research Centre, Institute of Computing Technology, Chinese Academy of Sciences. Her current research interests include federated learning, mobile edge computing and edge intelligence.

References (34)

  • WangL. et al.

    CMFL: Mitigating communication overhead for federated learning

  • NiknamS. et al.

    Federated learning for wireless communications: Motivation, opportunities and challenges

    (2019)
  • LimW.Y. et al.

    Federated learning in mobile edge networks: A comprehensive survey

    IEEE Commun. Surv. Tutor.

    (2020)
  • FangB. et al.

    Nestdnn: Resource-aware multi-tenant on-device deep learning for continuous mobile vision

  • WangX. et al.

    In-edge AI: Intelligentizing mobile edge computing, caching and communication by federated learning

    IEEE Netw.

    (2019)
  • WangS. et al.

    Adaptive federated learning in resource constrained edge computing systems

    IEEE J. Sel. Areas Commun.

    (2019)
  • TranN.H. et al.

    Federated learning over wireless networks: Optimization model design and analysis

  • Cited by (18)

    • An efficient federated learning framework for multi-channeled mobile edge network with layered gradient compression

      2023, Computer Networks
      Citation Excerpt :

      Federated learning has emerged as an efficient solution to analyze and process distributed data for data-driven tasks (e.g., autonomous driving, virtual reality, image classification, etc.) in Mobile Edge Computing (MEC) [1–6].

    • Federation based joint client and server side Machine Learning for 5G and beyond Wireless Channel Estimation

      2022, Computer Networks
      Citation Excerpt :

      The article provided a fundamental framework for FL and tested their proposed algorithms on public Google+ posts, where the authors analysed the frequency of different features of these posts as key indicators. The authors in [6], delineated that use of FL in edge intelligence technique can reduce transmission volume for data and improve the communication efficiency between end devices. The authors in [7] performed signal recognition using federation learning, where their focus is on elaborating the benefits of using federation in the future wireless network system in terms of data privacy, data size, data security, and high quality.

    • A survey on federated learning

      2021, Knowledge-Based Systems
      Citation Excerpt :

      The challenge of heterogeneity hinders federated learning. Some recent studies [83–86] have shown that if the heterogeneity of the devices in the system can be obtained in advance, all mobile devices can be grouped according to the heterogeneity, and a local central server can be assigned to each group. We can first aggregate a group of similarly heterogeneous device models, and then send them to the server to aggregate into a global model.

    View all citing articles on Scopus

    Hui Jiang received her B.S. degree from the Department of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing, China, in 2017. Since 2017, she is currently a Ph.D. candidate at the Networking Technology Research Centre, Institute of Computing Technology, Chinese Academy of Sciences. Her current research interests include federated learning, mobile edge computing and edge intelligence.

    Min Liu received her B.S. and M.S. degrees in computer science from Xi’an Jiaotong University, China, in 1999 and 2002, respectively. She got her Ph.D in computer science from the Graduate University of the Chinese Academy of Sciences in 2008. She is currently a professor at the Networking Technology Research Centre, Institute of Computing Technology, Chinese Academy of Sciences. Her current research interests include network coordination for unmanned system, wireless networks and mobile computing.

    Bo Yang received the Ph.D. degree in control science and engineering from the University of Chinese Academy of Sciences, Beijing, China, in 2017. He is currently a Research Associate with the Institute of Computing Technology, Chinese Academy of Sciences. His current research interests include wireless networks, edge computing, and multi-agent learning.

    Qingxiang Liu received his B.S. degree from Nanjing University of Information Science and Technology, Nanjing, China in 2020. Since 2020, he is currently a master candidate at the Networking Technology Research Centre, Institute of Computing Technology, Chinese Academy of Sciences. His current research interests include big data and mobile cloud computing.

    Jizhong Li is currently working in Huawei. His current research interest is focused on distributed computing and middleware.

    Xiaobing Guo received the B.S. and M.S. degrees in computer science from Xi’an Jiaotong University in 1999 and 2002, respectively, and the Ph.D. degree in computer science from the Graduate University of Chinese Academy of Sciences in 2013. He is currently a Principal Researcher in Lenovo Research. His current research interests include the blockchain and confidential computing.

    View full text