skip to main content
10.1145/1329125.1329174acmotherconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

Multiagent learning in adaptive dynamic systems

Published: 14 May 2007 Publication History

Abstract

Classically, an approach to the multiagent policy learning supposed that the agents, via interactions and/or by using preliminary knowledge about the reward functions of all players, would find an interdependent solution called "equilibrium". Recently, however, certain researchers question the necessity and the validity of the concept of equilibrium as the most important multiagent solution concept. They argue that a "good" learning algorithm is one that is efficient with respect to a certain class of counterparts. Adaptive players is an important class of agents that learn their policies separately from the maintenance of the beliefs about their counterparts' future actions and make their decisions based on that policy and the current belief. In this paper, we propose an efficient learning algorithm in presence of the adaptive counterparts called Adaptive Dynamics Learner (ADL), which is able to learn an efficient policy over the opponents' adaptive dynamics rather than over the simple actions and beliefs and, by so doing, to exploit these dynamics to obtain a higher utility than any equilibrium strategy can provide. We tested our algorithm on a substantial representative set of the most known and demonstrative matrix games and observed that ADL agent is highly efficient against Adaptive Play Q-learning (APQ) agent and Infinitesimal Gradient Ascent (IGA) agent. In self-play, when possible, ADL is able to converge to a Pareto optimal strategy maximizing the welfare of all players.

References

[1]
M. Bowling and M. Veloso. Multiagent learning using a variable learning rate. Artificial Intelligence, 136(2):215--250, 2002.
[2]
Y. Chang and L. Kaelbling. Playing is believing: The role of beliefs in multi-agent learning. In Proceedings of the Advances in Neural Information Processing Systems (NIPS'01), Canada, 2001.
[3]
C. Claus and C. Boutilier. The dynamics of reinforcement learning in cooperative multiagent systems. In Proceedings of the Fifteenth National Conference on Artificial Intelligence (AAAI'98), Menlo Park, CA, 1998. AAAI Press.
[4]
O. Gies and B. Chaib-draa. Apprentissage de la coordination multiagent: une méthode basée sur le Q-learning par jeu adaptatif. Revue d'Intelligence Artificielle, 20(2-3):385--412, 2006.
[5]
E. Nudelman, J. Wortman, K. Leyton-Brown, and Y. Shoham. Run the GAMUT: A comprehensive approach to evaluating game-theoretic algorithms. In Proceedings of Autonomous Agents and Multiagent Systems (AAMAS'04), 2004.
[6]
R. Powers and Y. Shoham. Learning against opponents with bounded memory. In Proceedings of the Nineteenth International Joint Conference on Artificial Intelligence (IJCAI'05), 2005.
[7]
R. Powers and Y. Shoham. New criteria and a new algorithm for learning in multi-agent systems. In L. K. Saul, Y. Weiss, and L. Bottou, editors, Advances in Neural Information Processing Systems, volume 17. MIT Press, 2005.
[8]
Y. Shoham, R. Powers, and T. Grenager. Multi-agent reinforcement learning: a critical survey. Technical report, Stanford University, 2003.
[9]
S. Singh, M. Kearns, and Y. Mansour. Nash convergence of gradient dynamics in general-sum games. In Proceedings of the Sixteenth Conference on Uncertainty in Artificial Intelligence (UAI'94), pages 541--548, San Francisco, CA, 1994. Morgan Kaufman.
[10]
G. Tesauro. Extending Q-learning to general adaptive multi-agent systems. In S. Thrun, L. Saul, and B. Scholkopf, editors, Advances in Neural Information Processing Systems, volume 16, Cambridge, MA, 2004. MIT Press.
[11]
H. Young. The evolution of conventions. Econometrica, 61(1):57--84, 1993.

Cited By

View all
  • (2019)The Gift Exchange GameProceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems10.5555/3306127.3331961(1913-1915)Online publication date: 8-May-2019
  • (2007)Competition and Coordination in Stochastic GamesProceedings of the 20th conference of the Canadian Society for Computational Studies of Intelligence on Advances in Artificial Intelligence10.1007/978-3-540-72665-4_3(26-37)Online publication date: 28-May-2007

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
AAMAS '07: Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
May 2007
1585 pages
ISBN:9788190426275
DOI:10.1145/1329125
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

  • IFAAMAS

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 May 2007

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. adaptation
  2. matrix games
  3. multiagent learning

Qualifiers

  • Research-article

Conference

AAMAS07
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,155 of 5,036 submissions, 23%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 05 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2019)The Gift Exchange GameProceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems10.5555/3306127.3331961(1913-1915)Online publication date: 8-May-2019
  • (2007)Competition and Coordination in Stochastic GamesProceedings of the 20th conference of the Canadian Society for Computational Studies of Intelligence on Advances in Artificial Intelligence10.1007/978-3-540-72665-4_3(26-37)Online publication date: 28-May-2007

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media