Abstract
This paper discusses the problem of territory division in Hide-and-Seek games. To obtain an efficient seeking performance for multiple seekers, the seekers should agree on searching their own territories and learn to visit good hiding places first so that the expected time to find the hider is minimized. We propose a learning model using Reinforcement Learning in a hierarchical learning structure. Elemental tasks of planning the path to each hiding place are learnt in the lower layer, and then the composite task of finding the optimal sequence is learnt in the higher layer. The proposed approach is examined on a set of different maps and resulted in convergece to the optimal solution.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Choset, H.: Coverage for robotics a survey of recent results. Annals of Mathematics and Artificial Intelligence 31(1-4), 113–126 (2001)
Agmon, N., Hazon, N., Kamink, G.A.: Constructing Spanning Trees for Efficient Multi-robot Coverage. In: Proceedings of the 2006 IEEE International Conference on Robotics and Automation, ICRA (2006)
Watkins, C.J.C.H., Dayan, P.: Q-learning. Machine Learning 8(3), 279–292 (1992)
Singh, S., Jaakkola, T., Littman, M.: Convergence results for single-step on-policy reinforcement-learning algorithms. Machine Learning (1998), 287–308 (2000)
Singh, S.P.: The Efficient Learning of Multiple Task Sequences. In: Advances in Neural Information Processing Systems, vol. 4, pp. 251–258 (1992)
Sutton, R.S.: Learning to predict by the methods of temporal differences. Machine Learning 3(1), 9–44 (1988)
Dijkstra, E.W.: A note on two problems in connexion with graphs. Numerische Mathematik 1, 269–271 (1959)
Gunady, M.K., Gomaa, W.: Reinforcement learning generalization using state aggregation with a maze-solving problem. In: Proceedings of IEEE Japan-Egypt Conference on Electronics, Communications and Computers (JEC-ECC), pp. 157–162 (2012)
Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: A survey. Arxiv preprint cs/9605103, pp. 237–285 (1996)
Mitchell, T.M.: Machine Learning book. McGraw-Hill Science/Engineering/Math (1997)
Watkins, C.J.C.H.: Learning from Delayed Rewards. PhD thesis, Cambridge Univ., Cambridge, England (1989)
Singh, S.P.: Transfer of learning by composing solutions for elemental sequential tasks. Machine Learning (1992)
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Gunady, M.K., Gomaa, W., Takeuchi, I. (2012). Multi-agent Task Division Learning in Hide-and-Seek Games. In: Ramsay, A., Agre, G. (eds) Artificial Intelligence: Methodology, Systems, and Applications. AIMSA 2012. Lecture Notes in Computer Science(), vol 7557. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33185-5_29
Download citation
DOI: https://doi.org/10.1007/978-3-642-33185-5_29
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-33184-8
Online ISBN: 978-3-642-33185-5
eBook Packages: Computer ScienceComputer Science (R0)