Skip to main content

Advertisement

Log in

Energy and Loss-aware Selective Updating for SplitFed Learning with Energy Harvesting-Powered Devices

  • Published:
Journal of Signal Processing Systems Aims and scope Submit manuscript

Abstract

SplitFed learning (SFL) is a promising data-privacy preserving decentralized learning framework for IoT devices that has low computation requirement but high communication overhead. To reduce the communication overhead, we present a selective model update method that sends/receives activations/gradients only in selected epochs. However for IoT devices that are powered by harvested energy, the client-side model computations can take place only when the harvested energy can support it. So in this paper, we propose an energy+loss-aware selective updating method for SFL systems where the client-side model is updated based on both the clients’ energy and loss changes. When all clients have the same energy harvesting capability, we show that the proposed method can save energy by 43.7% to 80.5% with 0.5% drop in accuracy compared to an energy-aware method for VGG11 and ResNet20 models on CIFAR-10 and CIFAR-100 datasets. When the energy harvesting capability of the clients are different, the proposed method can save energy by up to 28.8% to 70.0% with 0.5% drop in accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Figure 1
Figure 2
Figure 3
Figure 4
Figure 5
Figure 6
Figure 7
Figure 8
Figure 9
Figure 10
Figure 11
Figure 12
Figure 13

Similar content being viewed by others

Notes

  1. A synchronization system server is used here for clarity. In practice, the synchronization server and the server to compute server-side model are the same.

  2. We ignore the communication energy of transmitting client-side model for synchronization, since it is very small compared to others.

References

  1. Rachakonda, L., Mohanty, S. P., Kougianos, E., & Sundaravadivel, P. (2019). Stress-lysis: A DNN-integrated edge device for stress level detection in the IoMT. IEEE Transactions on Consumer Electronics, 65(4), 474–483.

    Article  Google Scholar 

  2. Ghenescu, V., Mihaescu, R. E., Carata, S. V., Ghenescu, M. T., Barnoviciu, E., & Chindea, M. (2018). Face detection and recognition based on general purpose DNN object detector. In 2018 International Symposium on Electronics and Telecommunications (ISETC) (pp. 1–4)IEEE.

  3. Chen, Y., Qin, X., Wang, J., Yu, C., & Gao, W. (2020). Fedhealth: A federated transfer learning framework for wearable healthcare. IEEE Intelligent Systems, 35(4), 83–93.

    Article  Google Scholar 

  4. Park, S., Kim, G., Kim, J., Kim, B., & Ye, J. C. (2021). Federated split vision transformer for COVID-19CXR diagnosis using task-agnostic training. arXiv preprint arXiv:2111.01338

  5. Liu, Y., James, J., Kang, J., Niyato, D., & Zhang, S. (2020). Privacy-preserving traffic flow prediction: A federated learning approach. IEEE Internet of Things Journal, 7(8), 7751–7763.

    Article  Google Scholar 

  6. Rocher, L., Hendrickx, J. M., & De Montjoye, Y. A. (2019). Estimating the success of re-identifications in incomplete datasets using generative models. Nature communications, 10(1), 1–9.

    Article  Google Scholar 

  7. McMahan, B., Moore, E., Ramage, D., Hampson, S., & Arcas, B. A. Y. (2017). Communication-efficient learning of deep networks from decentralized data. In Proceedings of Artificial Intelligence and Statistics (pp. 1273–1282). PMLR.

  8. Gupta, O., & Raskar, R. (2018). Distributed learning of deep neural network over multiple agents. Journal of Network and Computer Applications, 116, 1–8.

    Article  Google Scholar 

  9. Thapa, C., Chamikara, M. A .P., & Camtepe, S. (2020). Splitfed: When federated learning meets split learning. arXiv preprint arXiv:2004.12088

  10. Palanisamy, K., Khimani, V., Moti, M. H., & Chatzopoulos, D. (2021). Spliteasy: A practical approach for training ML models on mobile devices. In Proceedings of the 22nd International Workshop on Mobile Computing Systems and Applications (pp. 37–43).

  11. Bhat, G., Park, J., & Ogras, U. Y. (2017). Near-optimal energy allocation for self-powered wearable systems. In 2017 IEEE/ACM International Conference on Computer-Aided Design (ICCAD) (pp. 368–375).

  12. Sharma, H., Haque, A., & Jaffery, Z. A. (2019). Maximization of wireless sensor network lifetime using solar energy harvesting for smart agriculture monitoring. Ad Hoc Networks, 94, 101966.

  13. Boccalero, G., Boragno, C., Caviglia, D. D., & Morasso, R. (2016). Flehap: a wind powered supply for autonomous sensor nodes. Journal of Sensor and Actuator Networks, 5(4), 15.

    Article  Google Scholar 

  14. Chen, X., Li, J., & Chakrabarti, C. (2021). Communication and computation reduction for split learning using asynchronous training. In 2021 IEEE Workshop on Signal Processing Systems (SiPS) (pp. 76–81).

  15. Gao, Y., Kim, M., Abuadbba, S., Kim, Y., Thapa, C., Kim, K., Camtep, S. A., Kim, H., & Nepal, S. (2020). End-to-end evaluation of federated learning and split learning for internet of things. In 2020 International Symposium on Reliable Distributed Systems (SRDS) (pp. 91–100). IEEE.

  16. Singh, A., Vepakomma, P., Gupta, O. & Raskar, R. (2019). Detailed comparison of communication efficiency of split learning and federated learning. arXiv preprint arXiv:1909.09145

  17. Li, T., Sahu, A. K., Talwalkar, A., & Smith, V. (2020). Federated learning: Challenges, methods, and future directions. IEEE Signal Processing Magazine, 37(3), 50–60.

    Article  Google Scholar 

  18. Nori, M. K., Yun, S., & Kim, I.-M. (2021). Fast federated learning by balancing communication trade-offs. IEEE Transactions on Communications, 69(8), 5168–5182.

    Article  Google Scholar 

  19. Lin, Y., Han, S., Mao, H., Wang, Y., & Dally, B. (2018). Deep gradient compression: Reducing the communication bandwidth for distributed training. In Proceedings of International Conference on Learning Representations (ICLR) (pp. 1–14)

  20. Chen, Z., Xu, T.-B., Du, C., Liu, C.-L., & He, H. (2020). Dynamical channel pruning by conditional accuracy change for deep neural networks. IEEE Transactions on Neural Networks and Learning Systems, 32(2), 799–813.

    Article  Google Scholar 

  21. Diao, E., Ding, J., & Tarokh, V. (2021). HeteroFL: Computation and communication efficient federated learning for heterogeneous clients. In Proceedings of International Conference on Learning Representations, ICLR (pp. 1–24).

  22. Du, Y., Yang, S., & Huang, K. (2020). High-dimensional stochastic gradient quantization for communication-efficient edge learning. IEEE Transactions on Signal Processing, 68, 2128–2142.

    Article  MathSciNet  Google Scholar 

  23. Ko, J. H., Na, T., Amir, M. F., & Mukhopadhyay, S. (2018). Edge-host partitioning of deep neural networks with feature space encoding for resource-constrained internet-of-things platforms. In 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS) (pp. 1–6).

  24. Shi, W., Hou, Y., Zhou, S., Niu, Z., Zhang, Y., & Geng, L. (2019). Improving device-edge cooperative inference of deep learning via 2-step pruning. In IEEE INFOCOM 2019-IEEE Conference on Computer Communications Workshops (pp. 1–6). 

  25. Han, D., Bhatti, H. I., Lee, J., & Moon, J. (2021). Accelerating federated learning with split learning on locally generated losses. In ICML 2021 Workshop on Federated Learning for User Privacy and Data Confidentiality. ICML Board

  26. He, C., Annavaram, M. & Avestimehr, S. (2020). Group knowledge transfer: Federated learning of large CNNs at the edge. In Proceedings of Advances in Neural Information Processing Systems (NIPS) (vol. 33, pp. 14068–14080).

  27. Vepakomma, P., Gupta, O., Swedish, T., & Raskar, R. (2018). Split learning for health: Distributed deep learning without sharing raw patient data. arXiv preprint arXiv:1812.00564

  28. Abuadbba, S., Kim, K., Kim, M., Thapa, C., Camtepe, S. A., Gao, Y., Kim, H. & Nepal, S. (2020). Can we use split learning on 1D CNN models for privacy preserving training? In Proceedings of the 15th ACM Asia Conference on Computer and Communications Security (pp. 305–318).

  29. Güler, B., & Yener, A. (2021). A framework for sustainable federated learning. In 2021 19th International Symposium on Modeling and Optimization in Mobile, Ad hoc, and Wireless Networks (WiOpt) (pp. 1–8). IEEE.

  30. Hamdi, R., Chen, M., Said, A. B., Qaraqe, M., & Poor, H. V. (2021). Federated learning over energy harvesting wireless networks. IEEE Internet of Things Journal. https://doi.org/10.1109/JIOT.2021.3089054

    Article  Google Scholar 

  31. Pasquini, D., Ateniese, G., & Bernaschi, M. (2021). Unleashing the tiger: Inference attacks on split learning. In Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security (pp. 2113–2129).

  32. An, H., Venkatesan, S., Schiferl, S., Wesley, T., Zhang, Q., Wang, J., Choo, K., Liu, S., Liu, B., Li, Z. et al. (2020). A 170µw image signal processor enablinghierarchical image recognition for intelligence at the edge. In 2020 IEEE Symposium on VLSI Circuits (pp. 1–2).

  33. Kamath, S., & Lindh, J. (2010). Measuring Bluetooth low energy power consumption. Texas instruments application note AN092, Dallas.

  34. Mikhaylov, K., Stusek, M., Masek, P., Petrov, V., Petajajarvi, J., Andreev, S., Pokorny, J., Hosek, J., Pouttu, A., & Koucheryavy, Y. (2018). Multi-RAT LPWAN in smart cities: Trial of LoRaWAN and NB-IoT integration. In 2018 IEEE International Conference on Communications (ICC) (pp. 1–6).

  35. Sinha, R. S., Wei, Y., & Hwang, S. H. (2017). A survey on LPWA technology: LoRa and NB-IoT. ICT Express, 3(1), 14–21.

    Article  Google Scholar 

  36. Brock, A., Lim, T., Ritchie, J. M., & Weston, N. (2017). Freezeout: Accelerate training by progressively freezing layers. arXiv preprint arXiv:1706.04983

  37. Rajbhandari, S., Ruwase, O., Rasley, J., Smith, S. & He, Y. (2021). Zero-infinity: Breaking the GPU memory wall for extreme scale deep learning. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis (pp. 1–14).

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xing Chen.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

1.1 Energy Computation

Computation Energy

The overall client computation energy during training is a function of the number of epochs of computing forward and backward propagation:

$$\begin{aligned} E^{total}_{comp} = T_{act,k} \times E_{comp\_for} + T_{grad,k} \times E_{comp\_back} \end{aligned}$$
(8)

where \(T_{act,k}\) and \(T_{grad,k}\) is the number of epochs in which client k computes the forward propagation and backward propagation, respectively. \(E_{comp\_for}\) is the forward computation energy per epoch, and \(E_{comp\_back}\) is the backward computation energy per epoch. \(E_{comp\_for}\) and \(E_{comp\_back}\) are a function of the number of samples processed by client k, the number of operations in the forward and backward computations and the energy efficiency of the device:

$$\begin{aligned} E_{comp\_for} = n_k \times {OP_{comp\_for}} \times {e_{comp}} \end{aligned}$$
(9)
$$\begin{aligned} E_{comp\_back} = n_k \times {OP_{comp\_back}} \times {e_{comp}} \end{aligned}$$
(10)

where \(n_k\) is the number of samples in client k. \(OP_{comp\_for}\) and \(OP_{comp\_back}\) are the number of operations per sample necessary for forward and backward propagation. The computation operations of backward propagation is approximately twice that of forward propagation [37]. \(e_{comp}\) (Joule per operation) is the computation energy-efficiency of client-side devices.

Consider the SFL training on VGG11 with 3 convolution layers in client side with \(n_k= 2500\). The computation overhead of the client side model is \(OP_{comp\_for}=0.8\times 10^{-4}\) TOP/sample and \(OP_{comp\_back}=1.6\times 10^{-4}\) TOP/sample. When \(\tau _k=20\), the client-side model computes back propagation for 19 epochs, specifically in epochs 1-10, 21, 41,..., 181, thus \(T_{grad,k}=19\). The clients need to compute forward propagation for a total of 29 epochs, specifically in epochs 1-11, 21, 22, 41, 42,...,181, 182, thus \(T_{act,k}=29\). We assume the client-side devices is DNN accelerator with energy consumption of 1.5TOPs/W, that is 0.67 J/TOP [32]. As a result, \(E_{comp\_for}=0.13\) J, \(E_{comp\_back}=0.27\) J, \(E^{total}_{comp}=10\) J.

Communication Energy

The communication energy is a function of the number of epochs of sending activations and receiving gradients:

$$\begin{aligned} E^{total}_{comm} = T_{act,k} \times E_{comm\_{act}} + T_{grad,k} \times E_{comm\_{grad}} \end{aligned}$$
(11)

\(E_{comm\_{act}}\) and \(E_{comm\_{grad}}\) is the communication energy per epoch for sending activations and receiving gradients, respectively. And they are the function of the number of samples in client k, and the size of activations and gradients:

$$\begin{aligned} E_{comm\_{act}} = n_k \times act\_size \times {e_{TX}} \end{aligned}$$
(12)
$$\begin{aligned} E_{comm\_{grad}} = n_k \times grad\_size \times {e_{RX}} \end{aligned}$$
(13)

where \(T_{act,k}\) and \(T_{grad,k}\) is the number of epochs of client k to send activations and receive gradients, \(act\_size\) and \(grad\_size\) (bits) is the size in bits of activations and gradients per sample, and \(n_k\) is the number of samples in client k. \(e_{TX}\) and \(e_{RX}\) (J/bit) is the energy efficiency of transmitting and receiving data.

For the same example of SFL training on VGG11 with 3 convolution layers in client side with \(n_k= 2500\), the size of activations/gradients of client side model is \(256\times 8\times 8\times 32\text {bits}=0.5243\) Mbits per sample. When \(\tau _k=20\), \(T_{grad,k}=19\) and \(T_{act,k}=29\), as mentioned above. We calculate the energy consumption of three different communication protocols:

  • Bluetooth with 1Mbps@10mW [33]: \(e_{TX}\) = \(e_{RX}\) = \(10^{-8}\) J/bit, \(E_{comm\_{act}}=13.1\) J, \(E_{comm\_{grad}}=13.1\) J, \(E^{total}_{comm}=628\) J;

  • LoRaWan with 0.05Mbps@60mW [34]: \(e_{TX}\) = \(e_{RX}\) = \(1.2\times 10^{-6}\) J/bit, \(E_{comm\_{act}}=1573\) J, \(E_{comm\_{grad}}=1573\) J, \(E^{total}_{comm}=75504\) J;

  • NB-IoT with 0.25Mbps@200mW [35]: \(e_{TX}\) = \(e_{RX}\) = \(0.8\times 10^{-6}\) J/bit, \(E_{comm\_{act}}=1049\) J, \(E_{comm\_{grad}}=1049\) J, \(E^{total}_{comm}=50352\) J;

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chen, X., Li, J. & Chakrabarti, C. Energy and Loss-aware Selective Updating for SplitFed Learning with Energy Harvesting-Powered Devices. J Sign Process Syst 94, 961–975 (2022). https://doi.org/10.1007/s11265-022-01781-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11265-022-01781-4

Keywords

Navigation