skip to main content
10.1145/1160633.1160772acmconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
Article

Multi-agent reinforcement learning algorithm to handle beliefs of other agents' policies and embedded beliefs

Authors Info & Claims
Published:08 May 2006Publication History

ABSTRACT

We have developed a new series of multi-agent reinforcement learning algorithms that choose a policy based on beliefs about co-players' policies. The algorithms are applicable to situations where a state is fully observable by the agents, but there is no limit on the number of players. Some of the algorithms employ embedded beliefs to handle the cases that co-players are also choosing a policy based on their beliefs of others' policies. Simulation experiments on Iterated Prisoners' Dilemma games show that the algorithms using on policy-based belief converge to highly mutually-cooperative behavior, unlike the existing algorithms based on action-based belief.

References

  1. C. Claus and C. Boutilier. The dynamics of reinforcement learning in cooperative multiagent systems. In Proc. of AAAI-98, pages 746--752. 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. J. Hu and M. P. Wellman. Multiagent reinforcement learning: Theoretical framework and an algorithm. In Proc. of ICML 1998, pages 242--250. 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. M. L. Littman. Markov games as a framework for multi-agent reinforcement learning. In Proc. of ICML 1994, pages 157--163.1994.Google ScholarGoogle ScholarCross RefCross Ref
  4. T. Makino and K. Aihara. Self-observation principle for estimating the other's internal state. Mathematical Engineering Technical Reports METR 2003--36, the University of Tokyo, Oct. 2003.Google ScholarGoogle Scholar
  5. L. Panait and S. Luke. Cooperative multi-agent learning: The state of the art. Autonomous Agents and Multi-Agent Systems, 11:387--434, 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Y. Shoham, R. Powers, and T. Grenager. On the agenda(s) of research on multi-agent learning. In Proc. of Artificial Multiagent Learning: Papers from the 2004 AAAI Fall Symposium, Technical Report FS-04-02. 2004.Google ScholarGoogle Scholar
  7. M. Weinberg and J. S. Rosenschein. Best-response multiagent learning in non-stationary environments. In Proc. of AAMAS'04, pages 506--513, 2004. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Conferences
    AAMAS '06: Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems
    May 2006
    1631 pages
    ISBN:1595933034
    DOI:10.1145/1160633

    Copyright © 2006 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 8 May 2006

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • Article

    Acceptance Rates

    Overall Acceptance Rate1,155of5,036submissions,23%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader