skip to main content
10.1145/1329125.1329172acmotherconferencesArticle/Chapter ViewAbstractPublication PagesaamasConference Proceedingsconference-collections
research-article

Multiagent reinforcement learning and self-organization in a network of agents

Published:14 May 2007Publication History

ABSTRACT

To cope with large scale, agents are usually organized in a network such that an agent interacts only with its immediate neighbors in the network. Reinforcement learning techniques have been commonly used to optimize agents local policies in such a network because they require little domain knowledge and can be fully distributed. However, all of the previous work assumed the underlying network was fixed throughout the learning process. This assumption was important because the underlying network defines the learning context of each agent. In particular, the set of actions and the state space for each agent is defined in terms of the agent's neighbors. If agents dynamically change the underlying network structure (also called self-organize) during learning, then one needs a mechanism for transferring what agents have learned so far before (in the old network structure) to their new learning context (in the new network structure).

In this work we develop a novel self-organization mechanism that not only allows agents to self-organize the underlying network during the learning process, but also uses information from learning to guide the self-organization process. Consequently, our work is the first to study this interaction between learning and self-organization. Our self-organization mechanism uses heuristics to transfer the learned knowledge across the different steps of self-organization. We also present a more restricted version of our mechanism that is computationally less expensive and still achieve good performance. We use a simplified version of the distributed task allocation domain as our case study. Experimental results verify the stability of our approach and show a monotonic improvement in the performance of the learning process due to self-organization.

References

  1. S. Abdallah and V. Lesser. Learning the task allocation game. In Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS), 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. M. Bowling. Convergence and no-regret in multiagent learning. In Advances in Neural Information Processing Systems 17, pages 209--216. MIT Press, Cambridge, MA, 2005.Google ScholarGoogle Scholar
  3. M. Bowling and M. Veloso. Multiagent learning using a variable learning rate. Artificial Intelligence, 136(2):215--250, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. J. A. Boyan and M. L. Littman. Packet routing in dynamically changing networks: A reinforcement learning approach. In Advances in Neural Information Processing Systems, volume 6, pages 671--678. Morgan Kaufmann Publishers, Inc., 1994.Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. M. E. Gaston and M. desJardins. Agent-organized networks for dynamic team formation. In AAMAS '05: Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems, pages 230--237, New York, NY, USA, 2005. ACM Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. B. Horling. Quantitative Organizational Modeling and Design for Multi-Agent Systems. PhD thesis, University of Massachusetts at Amherst, February 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. L. Peshkin and V. Savova. Reinforcement learning for adaptive routing. In International Joint Conference on Neural Networks (IJCNN), pages 1825--1830, 2002.Google ScholarGoogle ScholarCross RefCross Ref
  8. M. Sims, C. Goldman, and V. Lesser. Self-organization through bottom-up coalition formation. In Proceedings of Second International Joint Conference on Autonomous Agents and MultiAgent Systems (AAMAS 2003), pages 867--874, Melbourne, AUS, July 2003. ACM Press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. M. E. Taylor and P. Stone. Behavior transfer for value-function-based reinforcement learning. In The Fourth International Joint Conference on Autonomous Agents and Multiagent Systems, pages 53--59, New York, NY, July 2005. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Multiagent reinforcement learning and self-organization in a network of agents

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        AAMAS '07: Proceedings of the 6th international joint conference on Autonomous agents and multiagent systems
        May 2007
        1585 pages
        ISBN:9788190426275
        DOI:10.1145/1329125

        Copyright © 2007 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 14 May 2007

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article

        Acceptance Rates

        Overall Acceptance Rate1,155of5,036submissions,23%

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader