Abstract
We present a pilot study focused on creating flexible Hierarchical Task Networks that can leverage Reinforcement Learning to repair and adapt incomplete plans in the simulated rich domain of Minecraft. This paper presents an early evaluation of our algorithm using simulation for adaptive agents planning in a dynamic world. Our algorithm uses an hierarchical planner and can theoretically be used for any type of “bot”. The main aim of our study is to create flexible knowledge-based planners for robots, which can leverage exploration and guide learning more efficiently by imparting structure using domain knowledge. Results from simulations indicate that a combined approach using both HTN and RL is more flexible than HTN alone and more efficient than RL alone.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Argall, B.D., Chernova, S., Veloso, M., Browning, B.: A survey of robot learning from demonstration. Robot. Auton. Syst. 57(5), 469–483 (2009). https://doi.org/10.1016/j.robot.2008.10.024
Becker, W.C.: Parents are Teachers: A Child Management Program (1971)
Belker, T., Hammel, M., Hertzberg, J.: Learning to optimize mobile robot navigation based on HTN plans. In: Proceedings of the IEEE International Conference on Robotics and Automation, ICRA 2003, vol. 3, pp. 4136–4141. IEEE (2003)
Berk, L.E., Winsler, A.: Scaffolding Children’s Learning: Vygotsky and Early Childhood Education. NAEYC Research into Practice Series, vol. 7. ERIC (1995)
Breazeal, C., Buchsbaum, D., Gray, J., Gatenby, D., Blumberg, B.: Learning from and about others: towards using imitation to bootstrap the social understanding of others by robots. Learning 11(1-2) (2006)
Breazeal, C., Scassellati, B.: Robots that imitate humans. Trends Cogn. Sci. 6(11), 481–487 (2002)
Cox, M.T., Raja, A.: Metareasoning: Thinking about Thinking. MIT Press, Cambridge (2011)
Erol, K., Hendler, J.A., Nau, D.S.: UMCP: a sound and complete procedure for hierarchical task-network planning. In: AIPS, vol. 94, pp. 249–254 (1994)
Erol, K., Hendler, J., Nau, D.S.: HTN planning: complexity and expressivity. In: AAAI, vol. 94, pp. 1123–1128 (1994)
Fitzgerald, T., Thomaz, A., Goel, A.: Human-Robot Co-Creativity: Task Transfer of a Spectrum of Similarity. In: Proceedings of Seventh International Conference on Computational Creativity, Atlanta, June 2017 (2017)
Hayashi, H., Tokura, S., Hasegawa, T., Ozaki, F.: Dynagent: an incremental forward-chaining HTN planning agent in dynamic domains. In: Baldoni, M., Endriss, U., Omicini, A., Torroni, P. (eds.) DALT 2005. LNCS (LNAI), vol. 3904, pp. 171–187. Springer, Heidelberg (2006). https://doi.org/10.1007/11691792_11
Hogg, C., Kuter, U., Muñoz-Avila, H.: Learning hierarchical task networks for nondeterministic planning domains. In: IJCAI International Joint Conference on Artificial Intelligence, pp. 1708–1714 (2009)
Jones, J.K., Goel, A.K.: Perceptually grounded self-diagnosis and self-repair of domain knowledge. Knowl. Based Syst. 27, 281–301 (2012)
Kober, J., Bagnell, J.A., Peters, J.: Reinforcement learning in robotics: a survey. Int. J. Robot. Res. 32(11), 1238–1278 (2013). http://repository.cmu.edu/robotics
Magnenat, S., Chappelier, J.-C., Mondada, F.: Integration of online learning into htn planning for robotic tasks. In: AAAI Spring Symposium: Designing Intelligent Robots (2012)
Magnenat, S., Voelkle, M., Mondada, F.: Planner9, a HTN planner distributed on groups of miniature mobile robots. In: Xie, M., Xiong, Y., Xiong, C., Liu, H., Hu, Z. (eds.) ICIRA 2009. LNCS (LNAI), vol. 5928, pp. 1013–1022. Springer, Heidelberg (2009). https://doi.org/10.1007/978-3-642-10817-4_99
Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D., Riedmiller, M.: Playing atari with deep reinforcement learning (2013). ArXiv preprint: arXiv:1312.5602
Mohseni-Kabir, A., Rich, C., Chernova, S., Sidner, C.L., Miller, D.: Interactive hierarchical task learning from a single demonstration. In: Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 205–212. ACM (2015)
Morimoto, J., Doya, K.: Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning. Robot. Auton. Syst. 36, 37–51 (2001)
Murdock, J.W., Goel, A.K.: Meta-case-based reasoning: self-improvement through self-understanding. J. Exp. Theor. Artif. Intell. 20(1), 1–36 (2008)
Nau, D., Au, T.-C., Ilghami, O., Kuter, U., Murdock, J.W., Wu, D., Yaman, F.: SHOP2: an HTN planning system. J. Artif. Intell. Res. 20, 379–404 (2003)
Nau, D., Cao, Y., Lotem, A., Munoz-Avila, H.: SHOP: simple hierarchical ordered planner (1999)
Nejati, N., Langley, P., Konik, T.: Learning hierarchical task networks by observation. In: Proceedings of the 23rd International Conference on Machine Learning, pp. 665–672. ACM (2006)
Sidowski, J.B., Wycko, L.B., Tabory, L.: The influence of reinforcement and punishment in a minimal social situation. J. Abnorm. Soc. Psychol. 52(1), 115 (1956)
Sun, R., Zhang, X.: Top-down versus bottom-up learning in cognitive skill acquisition. Cogn. Syst. Res. 5(1), 63–89 (2004)
Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction, vol. 1. MIT Press, Cambridge (1998)
Tessler, C., Givony, S., Zahavy, T., Mankowitz, D.J., Mannor, S.: A deep hierarchical approach to lifelong learning in minecraft, CoRR, volume abs/1604.07255 (2016). http://arxiv.org/abs/1604.07255
Ulam, P., Goel, A., Jones, J., Murdock, W.: Using model-based reflection to guide reinforcement learning. In: Reasoning, Representation, and Learning in Computer Games, p. 107 (2005)
Ulam, P., Jones, J., Goel, A.K.: Combining model-based meta-reasoning and reinforcement learning for adapting game-playing agents. In: Artificial Intelligence and Interactive Digital Entertainment Conference, pp. 132–137 (2008)
Williamson, M., Decker, K., Sycara, K.: Unified information and control flow in hierarchical task networks. In: Proceedings of the AAAI 1996 Workshop on Theories of Planning, Action, and Control (1996)
Wolfe, J., Marthi, B., Russell, S.: Combined task and motion planning for mobile manipulation. In: International Conference on Automated Planning and Scheduling, pp. 254–257 (2010)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Parashar, P., Sheneman, B., Goel, A.K. (2017). Adaptive Agents in Minecraft: A Hybrid Paradigm for Combining Domain Knowledge with Reinforcement Learning. In: Sukthankar, G., Rodriguez-Aguilar, J. (eds) Autonomous Agents and Multiagent Systems. AAMAS 2017. Lecture Notes in Computer Science(), vol 10643. Springer, Cham. https://doi.org/10.1007/978-3-319-71679-4_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-71679-4_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-71678-7
Online ISBN: 978-3-319-71679-4
eBook Packages: Computer ScienceComputer Science (R0)