Abstract
This paper proposes and evaluates MarLee, a multi-agent reinforcement learning system that integrates both exploitation- and exploration-oriented learning. Compared with conventional reinforcement learnings, MarLee is more robust in the face of a dynamically changing environment and is able to perform exploration-oriented learning efficiently even in a large-scale environment. Thus, MarLee is well suited for autonomous systems, for example, software agents and mobile robots, that operate in dynamic, large-scale environments, like the real-world and the Internet. Spreading activation, based on the behavior-based approach, is used to explore the environment, so by manipulating the parameters of the spreading activation, it is easy to tune the learning characteristics. The fundamental effectiveness of MarLee was demonstrated by simulation.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Barto, A.G., Bradtke, S.J., Singh, S.P.: Learning to act using real-time dynamic programing. Artificial Intelligence 72(1-2), 81–138 (1995)
Watkins, C.J.C., Dayan, P.: Technical Note: Q-Learning. Machine Learning 8, 55–68 (1992)
Grefenstette, J.J.: Credit Assignment in Rule Discovery Systems Based on Generic Algorithms. Machine Learning 3, 225–245 (1988)
Miyazaki, K., Yamamura, M., Kobayashi, S.: MarcoPolo: A Reinforcement Learning System Considering TradeOff Exploitation and Exploration under Markovian Environments. Journal of Japanese Society for Artificial Intelligence 12(1), 78–89 (1997) (in Japanese)
Minsky, M.: The Society of Mind. Simon & Schuster, New York (1986)
Holland, J.H., Reightman, J.S.: Cognitive Systems Based on Adaptive Algorithms. In: Waterman, D.A., Hayes-Roth, F. (eds.) Pattern-Directed Inference System. Academic Press, London (1978)
Maes, P.: The Agent Network Architecture (ANA). SIGART Bulletin 2(4), 115–120 (1991)
Maes, P.: Behavior-Based Artificial Intelligence. In: Proceedings of the Second International Conference on Simulation of Adaptive Behavior (From Animals to Animals 2), pp. 2–10. The MIT Press/Elsevier (1993)
Brooks, R.A.: A Robust Layered Control System for aMobile Robot. IEEE Robotics and Automation 2(1), 14–23 (1986)
Firby, R.J.: An Investigation into Reactive Planning in Complex Domains. In: AAAI 1987, pp. 189–208 (1987)
Kurihara, S., Onai, R.: Adaptive Selection of Reactive/Deliberate Planning for the Dynamic Environment. In: Proceedings of the MAAMAW 1997. LNCS (LNAI), vol. 1237 (1997)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 1998 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kurihara, S., Sugawara, T., Onai, R. (1998). Multi-Agent Reinforcement Learning System Integrating Exploitation- and Exploration-Oriented Learning. In: Zhang, C., Lukose, D. (eds) Multi-Agent Systems. Theories, Languages and Applications. DAI 1998. Lecture Notes in Computer Science(), vol 1544. Springer, Berlin, Heidelberg. https://doi.org/10.1007/10693067_4
Download citation
DOI: https://doi.org/10.1007/10693067_4
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-65477-3
Online ISBN: 978-3-540-49241-2
eBook Packages: Springer Book Archive