skip to main content
research-article

Decentralized Online Learning: Take Benefits from Others’ Data without Sharing Your Own to Track Global Trend

Published: 09 November 2022 Publication History

Abstract

Decentralized online learning (online learning in decentralized networks) has been attracting more and more attention, since it is believed that decentralized online learning can help data providers cooperatively better solve their online problems without sharing their private data to a third party or other providers. Typically, the cooperation is achieved by letting the data providers exchange their models between neighbors, e.g., recommendation model. However, the best regret bound for a decentralized online learning algorithm is 𝒪(nT), where n is the number of nodes (or users) and T is the number of iterations. This is clearly insignificant, since this bound can be achieved without any communication in the networks. This reminds us to ask a fundamental question: Can people really get benefit from the decentralized online learning by exchanging information? In this article, we studied when and why the communication can help the decentralized online learning to reduce the regret. Specifically, each loss function is characterized by two components: the adversarial component and the stochastic component. Under this characterization, we show that decentralized online gradient enjoys a regret bound \({\mathcal {O}(\sqrt {n^2TG^2 + n T \sigma ^2})}\), where G measures the magnitude of the adversarial component in the private data (or equivalently the local loss function) and σ measures the randomness within the private data. This regret suggests that people can get benefits from the randomness in the private data by exchanging private information. Another important contribution of this article is to consider the dynamic regret—a more practical regret to track users’ interest dynamics. Empirical studies are also conducted to validate our analysis.

References

[1]
Dmitry Adamskiy, Wouter M. Koolen, Alexey Chernov, and Vladimir Vovk. 2016. A closer look at adaptive regret. J. Mach. Learn. Res. 17, 23 (2016), 1–21.
[2]
M. Akbari, B. Gharesifard, and T. Linder. 2017. Distributed online convex optimization on time-varying directed graphs. IEEE Trans. Contr. Netw. Syst. 4, 3 (September 2017), 417–428.
[3]
A. S. Bedi, P. Sarma, and K. Rajawat. 2018. Tracking moving agents via inexact online gradient descent algorithm. IEEE J. Select. Top. Sign. Process. 12, 1 (February 2018), 202–217.
[4]
András A. Benczúr, Levente Kocsis, and Róbert Pálovics. 2018. Online machine learning in big data streams (unpublished).
[5]
Nicolò Cesa-Bianchi, Pierre Gaillard, Gabor Lugosi, and Gilles Stoltz. 2012. Mirror descent meets fixed share (and feels no regret). In Proceedings of the Conference and Workshop on Neural Information Processing Systems (NIPS’12). Paper 471.
[6]
Fanny Dufossé and Bora Uccar. 2016. Notes on Birkhoff-Von Neumann Decomposition of Doubly Stochastic Matrices. Research Report RR-8852. Inria–Research Centre Grenoble, Rhône-Alpes.
[7]
András György, Tamás Linder, and Gábor Lugosi. 2005. Tracking the best of many experts. InProceedings of Conference on Learning Theory (COLT’05).
[8]
A. Gyorgy, T. Linder, and G. Lugosi. 2012. Efficient tracking of large classes of experts. IEEE Trans. Inf. Theory 58, 11 (November 2012), 6709–6725.
[9]
András György and Csaba Szepesvári. 2016. Shifting regret, mirror descent, and matrices. In Proceedings of the 33rd International Conference on International Conference on Machine Learning, Volume 48 (ICML’16). 2943–2951.
[10]
Eric C. Hall and Rebecca Willett. 2013. Dynamical models and tracking regret in online convex programming. In Proceedings of International Conference on International Conference on Machine Learning (ICML’13).
[11]
Eric C. Hall and Rebecca M. Willett. 2015. Online convex optimization in dynamic environments.IEEE J. Select. Top. Sign. Process. 9, 4 (2015), 647–662.
[12]
Elad Hazan. 2016. Introduction to online convex optimization. Found. Trends Optim. 2, 3-4 (2016), 157–325.
[13]
Mark Herbster and Manfred K. Warmuth. 1998. Tracking the best expert. Mach. Learn. 32, 2 (August 1998), 151–178.
[14]
S. Hosseini, A. Chapman, and M. Mesbahi. 2013. Online distributed optimization via dual averaging. In Proceedings of the 52nd IEEE Conference on Decision and Control. 1484–1489.
[15]
Ali Jadbabaie, Alexander Rakhlin, Shahin Shahrampour, and Karthik Sridharan. 2015. Online optimization : Competing with dynamic comparators. In Proceedings of International Conference on Artificial Intelligence and Statistics (AISTATS’15). 398–406.
[16]
Jiyan Jiang, Wenpeng Zhang, Jinjie Gu, and Wenwu Zhu. 2021. Asynchronous decentralized online learning. In Advances in Neural Information Processing Systems. 20185–20196.
[17]
Kwang-Sung Jun, Francesco Orabona, Stephen Wright, and Rebecca Willett. 2017. Improved strongly adaptive online learning using coin betting. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS’17), Aarti Singh and Jerry Zhu (Eds.), Vol. 54. 943–951.
[18]
Michael Kamp, Mario Boley, Daniel Keren, Assaf Schuster, and Izchak Sharfman. 2014. Communication-efficient distributed online prediction by dynamic model synchronization. In Proceedings of the 2014th European Conference on Machine Learning and Knowledge Discovery in Databases, Volume Part I (ECMLPKDD’14). Springer-Verlag, Berlin, 623–639.
[19]
Ioannis Katakis, Grigorios Tsoumakas, and Ioannis Vlahavas. 2010. Tracking recurring contexts using ensemble classifiers: An application to email filtering. Knowl. Inf. Syst. 22, 3 (2010), 371–391.
[20]
P. Knight. 2008. The sinkhorn-knopp algorithm: Convergence and applications. SIAM J. Matrix Anal. Appl. 30, 1 (2008), 261–275.
[21]
A. Koppel, S. Paternain, C. Richard, and A. Ribeiro. 2018. Decentralized online learning with kernels. IEEE Trans. Sign. Process. 66, 12 (June 2018), 3240–3255.
[22]
S. Lee, A. Nedić, and M. Raginsky. 2018. Coordinate dual averaging for decentralized online optimization with nonseparable global objectives. IEEE Trans. Contr. Netw. Syst. 5, 1 (March 2018), 34–44.
[23]
S. Lee, A. Ribeiro, and M. M. Zavlanos. 2016. Distributed continuous-time online optimization using saddle-point methods. In Proceedings of the IEEE 55th Conference on Decision and Control (CDC’16). 4314–4319.
[24]
Mehryar Mohri and Scott Yang. 2018. Competing with automata-based expert sequences. In Proceedings of the 21st International Conference on Artificial Intelligence and Statistics, Amos Storkey and Fernando Perez-Cruz (Eds.), Vol. 84. 1732–1740.
[25]
Aryan Mokhtari, Shahin Shahrampour, Ali Jadbabaie, and Alejandro Ribeiro. 2016. Online optimization in dynamic environments: Improved regret rates for strongly convex problems. In Proceedings of IEEE Conference on Decision and Control (CDC’16). IEEE, 7195–7201.
[26]
Jaouad Mourtada and Odalric-Ambrym Maillard. 2017. Efficient tracking of a growing number of experts. Proceedings of the 28th International Conference on Algorithmic Learning Theory, Vol. 76, Proceedings of Machine Learning Research, 517–539.
[27]
A. Nedić, S. Lee, and M. Raginsky. 2015. Decentralized online optimization with global objectives and local communication. In Proceedings of the American Control Conference (ACC’15). 4497–4503.
[28]
Kevin Scaman, Francis Bach, Sébastien Bubeck, Yin Tat Lee, and Laurent Massoulié. 2018. Optimal algorithms for non-smooth distributed optimization in networks. In Proceedings of the 32nd International Conference on Neural Information Processing Systems. 2745–2754.
[29]
Flore Sentenac, Etienne Boursier, and Vianney Perchet. 2021. Decentralized learning in online queuing systems. In Advances in Neural Information Processing Systems, M. Ranzato, A. Beygelzimer, Y. Dauphin, P. S. Liang, and J. Wortman Vaughan (Eds.), Vol. 34. Curran Associates, Inc., 18501–18512.
[30]
S. Shahrampour and A. Jadbabaie. 2018. Distributed online optimization in dynamic environments using mirror descent. IEEE Trans. Automat. Contr. 63, 3 (March 2018), 714–725.
[31]
Shai Shalev-Shwartz. 2012. Online learning and online convex optimization. Found. Trends Mach. Learn. 4, 2 (2012), 107–194.
[32]
Hanlin Tang, Shaoduo Gan, Ce Zhang, Tong Zhang, and Ji Liu. 2018. Communication compression for decentralized training. arxiv:cs.LG/1803.06443v3. Retrieved from https://arxiv.org/abs/1803.06443v3.
[33]
Chen-Yu Wei, Yi-Te Hong, and Chi-Jen Lu. 2016. Tracking the best expert in non-stationary stochastic environments. In Proceedings of Advances in Neural Information Processing Systems, D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett (Eds.). 3972–3980.
[34]
T. Wu, K. Yuan, Q. Ling, W. Yin, and A. H. Sayed. 2018. Decentralized consensus optimization with asynchrony and delays. IEEE Trans. Sign. Inf. Process. Netw. 4, 2 (June 2018), 293–307.
[35]
Hao-Feng Xu, Qing Ling, and Alejandro Ribeiro. 2015. Online learning over a decentralized network through ADMM. J. Operat. Res. Soc. Chin. 3, 4 (December 2015), 537–562.
[36]
Yemao Xu, Dezun Dong, Yawei Zhao, Weixia Xu, and Xiangke Liao. 2020. OD-SGD: One-step delay stochastic gradient descent for distributed training. ACM Trans. Archit. Code Optim. 17, 4, Article 30 (September 2020), 26 pages.
[37]
F. Yan, S. Sundaram, S. V. N. Vishwanathan, and Y. Qi. 2013. Distributed autonomous online learning: Regrets and intrinsic privacy-preserving properties. IEEE Trans. Knowl. Data Eng. 25, 11 (November 2013), 2483–2493.
[38]
Tianbao Yang, Lijun Zhang, Rong Jin, and Jinfeng Yi. 2016. Tracking slowly moving clairvoyant - optimal dynamic regret of online learning with true and noisy gradient. In Proceedings of the 34th International Conference on Machine Learning (ICML’16).
[39]
Kun Yuan, Qing Ling, and Wotao Yin. 2016. On the convergence of decentralized gradient descent.SIAM Journal on Optimization 26, 3 (2016), 1835–1854.
[40]
J. Zeng and W. Yin. 2018. On nonconvex decentralized gradient descent. IEEE Trans. Sign. Process. 66, 11 (June 2018), 2834–2848.
[41]
Chi Zhang, Peilin Zhao, Shuji Hao, Yeng Chai Soh, Bu Sung Lee, Chunyan Miao, and Steven C. H. Hoi. 2018. Distributed multi-task classification: A decentralized online learning approach. Mach. Learn. 107, 4 (April 2018), 727–747.
[42]
Lijun Zhang, Shiyin Lu, and Zhi-Hua Zhou. 2018. Adaptive online learning in dynamic environments. In Advances in Neural Information Processing Systems 31, S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (Eds.). 1323–1333.
[43]
Lijun Zhang, Tianbao Yang, rong jin, and Zhi-Hua Zhou. 2018. Dynamic regret of strongly adaptive methods. In Proceedings of the 35th International Conference on Machine Learning (ICML’18). 5882–5891.
[44]
Lijun Zhang, Tianbao Yang, Jinfeng Yi, Rong Jin, and Zhi-Hua Zhou. 2017. Improved dynamic regret for non-degenerate functions. In Proceedings of the Annual Conference on Neural Information Processing Systems (NIPS’17).
[45]
Wenpeng Zhang, Peilin Zhao, Wenwu Zhu, Steven C. H. Hoi, and Tong Zhang. 2017. Projection-free distributed online learning in networks. In Proceedings of the 34th International Conference on Machine Learning, Doina Precup and Yee Whye Teh (Eds.). International Convention Centre, Sydney, Australia, 4054–4062.
[46]
Yawei Zhao, Shuang Qiu, Kuan Li, Lailong Luo, Jianping Yin, and Ji Liu. 2021. Proximal online gradient is optimum for dynamic regret: A general lower bound. IEEE Transactions on Neural Networks and Learning Systems, 1–10.
[47]
Yawei Zhao, Qian Zhao, Xingxing Zhang, En Zhu, Xinwang Liu, and Jianping Yin. 2020. Understand dynamic regret with switching cost for online decision making. ACM Trans. Intell. Syst. Technol. 11, 3 (April 2020), 21.
[48]
Martin Zinkevich. 2003. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of International Conference on Machine Learning (ICML’03). 928–935.

Cited By

View all
  • (2024)Optimized Gradient Tracking for Decentralized Online LearningIEEE Transactions on Signal Processing10.1109/TSP.2024.336643772(1443-1459)Online publication date: 2024
  • (2022)DAdam: A Consensus-Based Distributed Adaptive Gradient Method for Online OptimizationIEEE Transactions on Signal Processing10.1109/TSP.2022.322321470(6065-6079)Online publication date: 2022

Index Terms

  1. Decentralized Online Learning: Take Benefits from Others’ Data without Sharing Your Own to Track Global Trend

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Intelligent Systems and Technology
    ACM Transactions on Intelligent Systems and Technology  Volume 14, Issue 1
    February 2023
    487 pages
    ISSN:2157-6904
    EISSN:2157-6912
    DOI:10.1145/3570136
    • Editor:
    • Huan Liu
    Issue’s Table of Contents

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 09 November 2022
    Online AM: 02 September 2022
    Accepted: 01 August 2022
    Revised: 06 June 2022
    Received: 28 October 2020
    Published in TIST Volume 14, Issue 1

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. Decentralized online learning
    2. dynamic regret
    3. online gradient descent

    Qualifiers

    • Research-article
    • Refereed

    Funding Sources

    • Ministry of Industry and Information Technology of the People’s Republic of China
    • National Natural Science Foundation of China
    • National University of Defense Technology Foundation

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)100
    • Downloads (Last 6 weeks)9
    Reflects downloads up to 02 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Optimized Gradient Tracking for Decentralized Online LearningIEEE Transactions on Signal Processing10.1109/TSP.2024.336643772(1443-1459)Online publication date: 2024
    • (2022)DAdam: A Consensus-Based Distributed Adaptive Gradient Method for Online OptimizationIEEE Transactions on Signal Processing10.1109/TSP.2022.322321470(6065-6079)Online publication date: 2022

    View Options

    Login options

    Full Access

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Full Text

    View this article in Full Text.

    Full Text

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media