skip to main content
10.1145/1569901.1570034acmconferencesArticle/Chapter ViewAbstractPublication PagesgeccoConference Proceedingsconference-collections
research-article

Evolving an autonomous agent for non-Markovian reinforcement learning

Published:08 July 2009Publication History

ABSTRACT

In this paper, we investigate the use of nested evolution in which each step of one evolutionary process involves running a second evolutionary process. We apply this approach to build an evolutionary system for reinforcement learning (RL) problems. Genetic programming based on a descriptive encoding is used to evolve the neural architecture, while an evolution strategy is used to evolve the connection weights. We test this method on a non-Markovian RL problem involving an autonomous foraging agent, finding that the evolved networks significantly outperform a rule-based agent serving as a control. We also demonstrate that nested evolution, partitioning into subpopulations, and crossover operations all act synergistically in improving performance in this context.

References

  1. H. G. Beyer. The Theory of Evolution Strategies. Springer, Berlin, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. H. G. Beyer and H. P. Schwefel. Evolution strategies - a comprehensive introduction. Natural Computing, 1:3--52, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. W. H. Hsu and S. M. Gustafson. Genetic programming and multi-agent layered learning by reinforcements. In Proceedings of the Genetic and Evolutionary Computation Conference, pages 764--771, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. C. Igel. Neuroevolution for reinforcement learning using evolution strategies. Congress on Evolutionary Computation (CEC), volume 4, pages 2588--2595, IEEE Press, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  5. T. Jansen and I. Wegener. On the local performance of simulated annealing and the (1+1) evolutionary algorithm. In Proceedings of the 8th Annual Conference on Genetic and Evolutionary Computation, pages 469--476, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. J. Y. Jung and J. A. Reggia. Nested evolution of an autonomous agent using descriptive encoding. In Proceedings of the 10th Annual Conference on Genetic and Evolutionary Computation, pages 285--286, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. J. Y. Jung and J. A. Reggia. Evolutionary design of neural network architectures using a descriptive encoding language. IEEE Transactions on Evolutionary Computation, 10(6):676--688, Dec. 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Y. Kassahun and G. Sommer. Efficient reinforcement learning through evolutionary acquisition of neural topologies. In European Symposium on Artificial Neural Networks (ESANN), pages 259--266, d-side, 2005.Google ScholarGoogle Scholar
  9. W. N. Martin, J. Lienig, and J. P. Cohoon. Island (migration) models: evolutionary algorithms based on punctuated equilibria. In T. Bäck et al., editors, Handbook of Evolutionary Computation, pages 101--124, Institute of Physics Publishing and Oxford University Press, 1997.Google ScholarGoogle Scholar
  10. J. H. Metzen, M. Edgington, Y. Kassahun, and F. Kirchner. Analysis of an evolutionary reinforcement learning method in a multiagent domain. In Proceedings of the 7th international joint conference on Autonomous agents and multiagent systems, pages 291--298, 2008. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. F. Saibene and A. E. Minetti. Biomechanical and physiological aspects of legged locomotion in humans. European Journal of Applied Physiology, 88(4):297--316, 2003.Google ScholarGoogle ScholarCross RefCross Ref
  12. J. A. Reggia, S. Goodall, Y. Shkuro, and M. Glezer. The callosal dilemma: Explaining diaschisis in the context of hemispheric rivalry via a neural network model. Neurological Research, 23:465--471, 2001.Google ScholarGoogle ScholarCross RefCross Ref
  13. J. A. Reggia, R. Schulz, G. Wilkinson, and J. Uriagereka. Conditions enabling the evolution of inter-agent signaling in an artificial world. Artificial Life, 7(1):3--32, 2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. E. Ruppin. Evolutionary autonomous agents: A neuroscience perspective. Nature Reviews Neuroscience, 3(2):132--141, 2002.Google ScholarGoogle ScholarCross RefCross Ref
  15. N. Siebel and G. Sommer. Evolutionary reinforcement learning of artificial neural networks. International Journal of Hybrid Intelligent Systems, 4(3):171--183, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. K. O. Stanley and R. Miikkulainen. Evolving neural network through augmenting topologies. Evolutionary Computation, 10(2):99--127, 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. P. Stone, R. S. Sutton, and G. Kuhlmann. Reinforcement Learning for RoboCup-Soccer Keepaway. Adaptive Behavior, 13(3):165--188, 2005.Google ScholarGoogle ScholarCross RefCross Ref
  18. R. S. Sutton and A. G. Barto. Reinforcement Learning An Introduction. MIT Press, 1998. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. I. Szita and A. Lörincz. Learning tetris using the noisy cross-entropy method. Neural Computation 18(12): 2936--2941, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. M. E. Taylor, S. Whiteson, and P. Stone. Comparing evolutionary and temporal difference methods in a reinforcement learning domain. In Proceedings of the 8th Annual Conference on Genetic and Evolutionary Computation, pages 1321--1328, 2006. Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. S. Whiteson, M. E. Taylor, and P. Stone. Empirical studies in action selection for reinforcement learning. Adaptive Behavior, 15(1):33--50, 2007. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. X. Yao. Evolving artificial neural networks. Proceedings of the IEEE, 87(9):1423--1447, Sept. 1999.Google ScholarGoogle ScholarCross RefCross Ref
  23. X. Yao and Y. Liu. Evolutionary artificial neural networks that learn and generalize well. In Proceedings of the 1996 IEEE International Conference on Neural Networks, pages 159--164, 1996.Google ScholarGoogle Scholar

Index Terms

  1. Evolving an autonomous agent for non-Markovian reinforcement learning

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      GECCO '09: Proceedings of the 11th Annual conference on Genetic and evolutionary computation
      July 2009
      2036 pages
      ISBN:9781605583259
      DOI:10.1145/1569901

      Copyright © 2009 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 8 July 2009

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      Overall Acceptance Rate1,669of4,410submissions,38%

      Upcoming Conference

      GECCO '24
      Genetic and Evolutionary Computation Conference
      July 14 - 18, 2024
      Melbourne , VIC , Australia

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader