Skip to main content

Continuous-Time Multi-agent Network for Distributed Least Absolute Deviation

  • Conference paper
  • First Online:
Advances in Neural Networks – ISNN 2015 (ISNN 2015)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 9377))

Included in the following conference series:

Abstract

This paper presents a continuous-time multi-agent network for distributed least absolute deviation (DLAD). The objective function of the DLAD problem is a sum of many least absolute deviation functions. In the multi-agent network, each agent connects with its neighbors locally and they cooperate to obtain the optimal solutions with consensus. The proposed multi-agent network is in fact a collective system with each agent being considered as a recurrent neural network. Simulation results on a numerical example are presented to illustrate the effectiveness and characteristics of the proposed distributed optimization method.

This work was supported in part by the National Natural Science Foundation of China under Grant 61473333, by the Program for New Century Excellent Talents in University of China under Grant NCET-12-0114, and by the Fundamental Research Funds for the Central Universities of China under Grant 2015QN035.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Cochocki, A., Unbehauen, R.: Neural Networks for Optimization and Signal Processing. John Wiley & Sons, New York (1993)

    Google Scholar 

  2. Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning 3, 1–122 (2011)

    Article  MATH  Google Scholar 

  3. Nedic, A., Ozdaglar, A., Parrilo, P.A.: Constrained consensus and optimization in multi-agent networks. IEEE Transactions on Automatic Control 55, 922–938 (2010)

    Article  MathSciNet  MATH  Google Scholar 

  4. Zhu, M., Martínez, S.: On distributed convex optimization under inequality and equality constraints. IEEE Transactions on Automatic Control 57, 151–164 (2012)

    Article  MathSciNet  MATH  Google Scholar 

  5. Gharesifard, B., Cortés, J.: Distributed continuous-time convex optimization on weight-balanced digraphs. IEEE Transactions on Automatic Control 59, 781–786 (2014)

    Article  MathSciNet  MATH  Google Scholar 

  6. Lobel, I., Ozdaglar, A.: Distributed subgradient methods for convex optimization over random networks. IEEE Transactions on Automatic Control 56, 1291–1306 (2011)

    Article  MathSciNet  MATH  Google Scholar 

  7. Liu, Q., Wang, J.: A second-order multi-agent network for bound-constrained distributed optimization. IEEE Transactions on Automatic Control (2015), doi:10.1109/TAC.2015.2416927.

    Google Scholar 

  8. Lin, P., Jia, Y.: Consensus of a class of second-order multi-agent systems with time-delay and jointly-connected topologies. IEEE Transactions on Automatic Control 55, 778–784 (2010)

    Article  MathSciNet  MATH  Google Scholar 

  9. Mei, J., Ren, W., Ma, G.: Distributed coordination for second-order multi-agent systems with nonlinear dynamics using only relative position measurements. Automatica 49, 1419–1427 (2013)

    Article  MathSciNet  MATH  Google Scholar 

  10. Yu, W., Ren, W., Zheng, W.X., Chen, G., Lü, J.: Distributed control gains design for consensus in multi-agent systems with second-order nonlinear dynamics. Automatica 49, 2107–2115 (2013)

    Article  MathSciNet  MATH  Google Scholar 

  11. Tank, D., Hopfield, J.: Simple neural optimization networks: An a/d converter, signal decision circuit, and a linear programming circuit. IEEE Transactions on Circuits and Systems 33, 533–541 (1986)

    Article  Google Scholar 

  12. Kennedy, M., Chua, L.: Neural networks for nonlinear programming. IEEE Transactions on Circuits and Systems 35, 554–562 (1988)

    Article  MathSciNet  Google Scholar 

  13. Xia, Y., Wang, J.: A general projection neural network for solving monotone variational inequalities and related optimization problems. IEEE Transactions on Neural Networks 15, 318–328 (2004)

    Article  Google Scholar 

  14. Hu, X., Wang, J.: Solving the assignment problem using continuous-time and discrete-time improved dual networks. IEEE Transactions on Neural Networks and Learning Systems 23, 821–827 (2012)

    Article  Google Scholar 

  15. Liu, Q., Wang, J.: A one-layer projection neural network for nonsmooth optimization subject to linear equalities and bound constraints. IEEE Transactions on Neural Networks and Learning Systems 24, 812–824 (2013)

    Article  Google Scholar 

  16. Liu, Q., Dang, C., Huang, T.: A one-layer recurrent neural network for real-time portfolio optimization with probability criterion. IEEE Transactions on Cybernetics 43, 14–23 (2013)

    Article  Google Scholar 

  17. Liu, Q., Huang, T.: A neural network with a single recurrent unit for associative memories based on linear optimization. Neurocomputing 118, 263–267 (2013)

    Article  Google Scholar 

  18. Liu, Q., Huang, T., Wang, J.: One-layer continuous- and discrete-time projection neural networks for solving variational inequalities and related optimization problems. IEEE Transactions on Neural Networks and Learning Systems 25, 1308–1318 (2014)

    Article  Google Scholar 

  19. Liu, Q., Wang, J.: A projection neural network for constrained quadratic minimax optimization (2015), doi:10.1109/TNNLS.2015.2425301.

    Google Scholar 

  20. Nedic, A., Ozdaglar, A.: Distributed subgradient methods for multi-agent optimization. IEEE Transactions on Automatic Control 54, 48–61 (2009)

    Article  MathSciNet  MATH  Google Scholar 

  21. Kinderlehrer, D., Stampacchia, G.: An Introduction to Variational Inequalities and Their Applications. Academic, New York (1982)

    MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Qingshan Liu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Liu, Q., Zhao, Y., Cheng, L. (2015). Continuous-Time Multi-agent Network for Distributed Least Absolute Deviation. In: Hu, X., Xia, Y., Zhang, Y., Zhao, D. (eds) Advances in Neural Networks – ISNN 2015. ISNN 2015. Lecture Notes in Computer Science(), vol 9377. Springer, Cham. https://doi.org/10.1007/978-3-319-25393-0_48

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-25393-0_48

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-25392-3

  • Online ISBN: 978-3-319-25393-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics