Skip to main content

Combining Planning with Reinforcement Learning for Multi-robot Task Allocation

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 3394))

Abstract

We describe an approach to the multi-robot task allocation (MRTA) problem in which a group of robots must perform tasks that arise continuously, at arbitrary locations across a large space. A dynamic scheduling algorithm is derived in which proposed plans are evaluated using a combination of short-term lookahead and a value function acquired by reinforcement learning. We demonstrate that this dynamic scheduler can learn not only to allocate robots to tasks efficiently, but also to position the robots appropriately in readiness for new tasks (tactical awareness), and conserve resources over the long run (strategic awareness).

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Boutilier, C., Dean, T., Hanks, S.: Decision-theoretic planning: Structural assump- tions and computational leverage. Journal of Artificial Intelligence Research 11, 1–94 (1999)

    MATH  MathSciNet  Google Scholar 

  2. Gerkey, B.P., Mataric, M.J.: A formal framework for study of task allocation in multi-robot systems. Technical Report CRES-03-13, University of Southern Cali-fornia (2003)

    Google Scholar 

  3. Bellman, R.E.: Dynamic Programming. Princeton University Press, Princeton (1957)

    MATH  Google Scholar 

  4. Sutton, R.S., Barto, A.G.: Reinforcement Learning. MIT Press, Cambridge (1998)

    Google Scholar 

  5. Watkins, C.J.C.H.: Models of Delayed Reinforcement Learning. PhD thesis, Psy- chology Department, Cambridge University, Cambridge, United Kingdom (1989)

    Google Scholar 

  6. Strens, M.J.A., Moore, A.W.: Policy search using paired comparisons. Journal of Machine Learning Research 3, 921–950 (2002)

    Article  Google Scholar 

  7. Martin, J.J.: Bayesian Decision problems and Markov Chains. John Wiley, New York (1967)

    MATH  Google Scholar 

  8. Meuleau, N., Hauskrecht, M., Kim, K.E., Peshkin, L., Kaelbling, L.P., Dean, T., Boutilier, C.: Solving very large weakly coupled Markov decision processes. In: Proceedings of the 15th National Conference on Artificial Intelligence (AAAI 1998), pp. 165–172. AAAI Press, Menlo Park (1998)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Strens, M., Windelinckx, N. (2005). Combining Planning with Reinforcement Learning for Multi-robot Task Allocation. In: Kudenko, D., Kazakov, D., Alonso, E. (eds) Adaptive Agents and Multi-Agent Systems II. AAMAS AAMAS 2004 2003. Lecture Notes in Computer Science(), vol 3394. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-32274-0_17

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-32274-0_17

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-25260-3

  • Online ISBN: 978-3-540-32274-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics