Skip to main content

Multi-agent Task Division Learning in Hide-and-Seek Games

  • Conference paper
Artificial Intelligence: Methodology, Systems, and Applications (AIMSA 2012)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 7557))

  • 1064 Accesses

Abstract

This paper discusses the problem of territory division in Hide-and-Seek games. To obtain an efficient seeking performance for multiple seekers, the seekers should agree on searching their own territories and learn to visit good hiding places first so that the expected time to find the hider is minimized. We propose a learning model using Reinforcement Learning in a hierarchical learning structure. Elemental tasks of planning the path to each hiding place are learnt in the lower layer, and then the composite task of finding the optimal sequence is learnt in the higher layer. The proposed approach is examined on a set of different maps and resulted in convergece to the optimal solution.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Choset, H.: Coverage for robotics a survey of recent results. Annals of Mathematics and Artificial Intelligence 31(1-4), 113–126 (2001)

    Article  Google Scholar 

  2. Agmon, N., Hazon, N., Kamink, G.A.: Constructing Spanning Trees for Efficient Multi-robot Coverage. In: Proceedings of the 2006 IEEE International Conference on Robotics and Automation, ICRA (2006)

    Google Scholar 

  3. Watkins, C.J.C.H., Dayan, P.: Q-learning. Machine Learning 8(3), 279–292 (1992)

    MATH  Google Scholar 

  4. Singh, S., Jaakkola, T., Littman, M.: Convergence results for single-step on-policy reinforcement-learning algorithms. Machine Learning (1998), 287–308 (2000)

    Article  Google Scholar 

  5. Singh, S.P.: The Efficient Learning of Multiple Task Sequences. In: Advances in Neural Information Processing Systems, vol. 4, pp. 251–258 (1992)

    Google Scholar 

  6. Sutton, R.S.: Learning to predict by the methods of temporal differences. Machine Learning 3(1), 9–44 (1988)

    Google Scholar 

  7. Dijkstra, E.W.: A note on two problems in connexion with graphs. Numerische Mathematik 1, 269–271 (1959)

    Article  MathSciNet  MATH  Google Scholar 

  8. Gunady, M.K., Gomaa, W.: Reinforcement learning generalization using state aggregation with a maze-solving problem. In: Proceedings of IEEE Japan-Egypt Conference on Electronics, Communications and Computers (JEC-ECC), pp. 157–162 (2012)

    Google Scholar 

  9. Kaelbling, L.P., Littman, M.L., Moore, A.W.: Reinforcement learning: A survey. Arxiv preprint cs/9605103, pp. 237–285 (1996)

    Google Scholar 

  10. Mitchell, T.M.: Machine Learning book. McGraw-Hill Science/Engineering/Math (1997)

    Google Scholar 

  11. Watkins, C.J.C.H.: Learning from Delayed Rewards. PhD thesis, Cambridge Univ., Cambridge, England (1989)

    Google Scholar 

  12. Singh, S.P.: Transfer of learning by composing solutions for elemental sequential tasks. Machine Learning (1992)

    Google Scholar 

  13. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Gunady, M.K., Gomaa, W., Takeuchi, I. (2012). Multi-agent Task Division Learning in Hide-and-Seek Games. In: Ramsay, A., Agre, G. (eds) Artificial Intelligence: Methodology, Systems, and Applications. AIMSA 2012. Lecture Notes in Computer Science(), vol 7557. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-33185-5_29

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-33185-5_29

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-33184-8

  • Online ISBN: 978-3-642-33185-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics