skip to main content
10.1145/3520304.3529073acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
poster

Using evolutionary game theory to understand scalability in task allocation

Published: 19 July 2022 Publication History

Abstract

Cooperation is an important challenge in multi-agent systems. Decentralised learning of cooperation is difficult because interactions between agents make the environment non-stationary, and the reward structure tempts agents to act selfishly. A centralised solution bypasses these challenges, but may scale poorly with system size. Understanding this trade-off is important, but systematic comparisons have been limited to tasks with fully aligned incentives. We introduce a new task for studying cooperation: agents can solve the task by working together and specialising in different sub-tasks, or by working alone. Using neuroevolution, we empirically investigate scalability comparing centralised and decentralised approaches. A mathematical model based on the replicator dynamics allows us to further study how the task's social dynamics affect the emergent behaviour. Our results show that the task's unique social features, in particular the challenge of agents' physical coordination, causes both centralised and decentralised approaches to scale poorly. We conclude that mitigating this coordination challenge can improve scalability more than the choice of learning type.

Supplementary Material

ZIP File (p152-rizk-suppl.zip)
Supplemental material.

References

[1]
Nicolas Anastassacos, Julian García, Stephen Hailes, and Mirco Musolesi. 2021. Cooperation and Reputation Dynamics with Reinforcement Learning. In Proceedings of the 20th International Conference on Autonomous Agents and MultiAgent Systems (AAMAS '21). International Foundation for Autonomous Agents and Multiagent Systems, Richland, SC, 115--123.
[2]
Bowen Baker, Ingmar Kanitscheider, Todor Markov, Yi Wu, Glenn Powell, Bob McGrew, and Igor Mordatch. 2019. Emergent tool use from multi-agent autocurricula. arXiv preprint arXiv:1909.07528 (2019).
[3]
Christopher Berner, Greg Brockman, Brooke Chan, Vicki Cheung, Przemysław Dębiak, Christy Dennison, David Farhi, Quirin Fischer, Shariq Hashme, Chris Hesse, et al. 2019. Dota 2 with large scale deep reinforcement learning. arXiv preprint arXiv:1912.06680 (2019).
[4]
Arne Brutschy, Giovanni Pini, Carlo Pinciroli, Mauro Birattari, and Marco Dorigo. 2014. Self-organized task allocation to sequentially interdependent tasks in swarm robotics. Autonomous agents and multi-agent systems 28, 1 (2014), 101--125.
[5]
Allan Dafoe, Yoram Bachrach, Gillian Hadfield, Eric Horvitz, Kate Larson, and Thore Graepel. 2021. Cooperative AI: Machines Must Learn to Find Common Ground. Nature 593, 7857 (May 2021), 33--36.
[6]
Eliseo Ferrante, Ali Emre Turgut, Edgar Duéñez-Guzmán, Marco Dorigo, and Tom Wenseleers. 2015. Evolution of self-organized task specialization in robot swarms. PLoS computational biology 11, 8 (2015), e1004273.
[7]
Jakob Foerster, Richard Y Chen, Maruan Al-Shedivat, Shimon Whiteson, Pieter Abbeel, and Igor Mordatch. 2018. Learning with Opponent-Learning Awareness. (2018), 9.
[8]
Herbert Gintis. 2009. Game theory evolving. Princeton university press.
[9]
Jorge Gomes, Pedro Mariano, and Anders Lyhne Christensen. 2017. Novelty-driven cooperative coevolution. Evolutionary computation 25, 2 (2017), 275--307.
[10]
Jayesh K Gupta, Maxim Egorov, and Mykel Kochenderfer. 2017. Cooperative multi-agent control using deep reinforcement learning. In International Conference on Autonomous Agents and Multiagent Systems. Springer, 66--83.
[11]
Nikolaus Hansen. 2016. The CMA evolution strategy: A tutorial. arXiv preprint arXiv:1604.00772 (2016).
[12]
A Jamshidpey, M Wahby, MK Heinrich, M Allwright, W Zhu, and M Dorigo. 2021. Centralization vs. decentralization in multi-robot coverage: Ground robots under UAV supervision. (2021).
[13]
Alaa Khamis, Ahmed Hussein, and Ahmed Elmogy. 2015. Multi-robot task allocation: A review of the state-of-the-art. Cooperative Robots and Sensor Networks 2015 (2015), 31--51.
[14]
Adam Lerer and Alexander Peysakhovich. 2017. Maintaining Cooperation in Complex Social Dilemmas Using Deep Reinforcement Learning. (July 2017).
[15]
Declan Oller, Tobias Glasmachers, and Giuseppe Cuccu. 2020. Analyzing Reinforcement Learning Benchmarks with Random Weight Guessing. In Proceedings of the 19th International Conference on Autonomous Agents and MultiAgent Systems. 975--982.
[16]
Liviu Panait and Sean Luke. 2005. Cooperative multi-agent learning: The state of the art. Autonomous agents and multi-agent systems 11, 3 (2005), 387--434.
[17]
Sven Seuken and Shlomo Zilberstein. 2007. Improved memory-bounded dynamic programming for decentralized POMDPs. In Proceedings of the Twenty-Third Conference on Uncertainty in Artificial Intelligence. 344--351.
[18]
Kenneth O Stanley, Jeff Clune, Joel Lehman, and Risto Miikkulainen. 2019. Designing neural networks through neuroevolution. Nature Machine Intelligence 1, 1 (2019), 24--35.
[19]
Felipe Petroski Such, Vashisht Madhavan, Edoardo Conti, Joel Lehman, Kenneth O Stanley, and Jeff Clune. 2017. Deep neuroevolution: Genetic algorithms are a competitive alternative for training deep neural networks for reinforcement learning. arXiv preprint arXiv:1712.06567 (2017).
[20]
Rinde RS van Lon and Tom Holvoet. 2017. When do agents outperform centralized algorithms? Autonomous Agents and Multi-Agent Systems 31, 6 (2017), 1578--1609.
[21]
Chern Han Yong and Risto Miikkulainen. 2009. Coevolution of role-based cooperation in multiagent systems. IEEE Transactions on Autonomous Mental Development 1, 3 (2009), 170--186.

Cited By

View all
  • (2024)Pflacco: Feature-Based Landscape Analysis of Continuous and Constrained Optimization Problems in PythonEvolutionary Computation10.1162/evco_a_00341(1-6)Online publication date: 15-Feb-2024
  • (2024)Evolutionary Game Theory as a Catalyst in Smart Grids: From Theoretical Insights to Practical StrategiesIEEE Access10.1109/ACCESS.2024.343693512(186926-186940)Online publication date: 2024

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
GECCO '22: Proceedings of the Genetic and Evolutionary Computation Conference Companion
July 2022
2395 pages
ISBN:9781450392686
DOI:10.1145/3520304
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the Owner/Author.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 19 July 2022

Check for updates

Author Tags

  1. co-evolution
  2. complex systems
  3. evolution strategies
  4. multi-agent systems
  5. neuroevolution

Qualifiers

  • Poster

Conference

GECCO '22
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,669 of 4,410 submissions, 38%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)29
  • Downloads (Last 6 weeks)1
Reflects downloads up to 02 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Pflacco: Feature-Based Landscape Analysis of Continuous and Constrained Optimization Problems in PythonEvolutionary Computation10.1162/evco_a_00341(1-6)Online publication date: 15-Feb-2024
  • (2024)Evolutionary Game Theory as a Catalyst in Smart Grids: From Theoretical Insights to Practical StrategiesIEEE Access10.1109/ACCESS.2024.343693512(186926-186940)Online publication date: 2024

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media