Skip to main content

Unique State and Automatical Action Abstracting Based on Logical MDPs with Negation

  • Conference paper
Advances in Natural Computation (ICNC 2006)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 4222))

Included in the following conference series:

  • 913 Accesses

Abstract

In this paper we introduce negation into Logical Markov Decision Processes, which is a model of Relational Reinforcement Learning. In the new model nLMDP the abstract state space can be constructed in a simple way, so that a good property of complementarity holds. Prototype action is also introduced into the model. A distinct feature of the model is that applicable abstract actions can be obtained automatically with valid substitutions. Given a complementary abstract state space and a set of prototype actions, a model-free Θ-learing method is implemented for evaluating the state-action-substitution value funcion.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Kaelbling, L., Littman, M., Moore, A.: Reinforcement learning: A survey. Journal of Artificial Intelligence Research 4, 237–285 (1996)

    Google Scholar 

  2. Sutton, R., Barto, A.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)

    Google Scholar 

  3. Džeroski, S., De Raedt, L., Driessens, K.: Relational reinforcement learning. Machine Learning 43, 7–52 (2001)

    Article  MATH  Google Scholar 

  4. Driessens, K., Ramon, J.: Relational instrance based regression for relational reinforcement learning. In: ICML 2003 (2003)

    Google Scholar 

  5. Gärtner, T., Driessens, K., Ramon, J.: Graph kernels and gaussian processes for relational reinforcement learning. In: Horváth, T., Yamamoto, A. (eds.) ILP 2003. LNCS, vol. 2835, pp. 146–163. Springer, Heidelberg (2003)

    Chapter  Google Scholar 

  6. Cole, J., Lloyd, K., Ng, K.: Symbolic learning for adaptive agents. In: The Annual Partner Conference, Smart Internet Technology Cooperative Research Centre (2003)

    Google Scholar 

  7. Yoon, S., Fern, A., Givan, R.: Inductive policy selection for first-order MDPs. In: UAI 2002 (2002)

    Google Scholar 

  8. Fern, A., Yoon, S., Givan, R.: Approximate policy iteration with a policy language bias. In: NIPS 2003 (2003)

    Google Scholar 

  9. Boutilier, C., Reiter, R., Price, B.: Symbolic dynamic programming for first-order MDPs. In: Seventeenth International Joint Conference on Artificial Intelligence (IJCAI 2001), pp. 690–700 (2001)

    Google Scholar 

  10. Kersting, K., De Raedt, L.: Logical markov decision programs. In: IJCAI 2003 Workshop on Learning Statistical Models of Relational Data (2003)

    Google Scholar 

  11. Van Otterlo, M.: Reinforcement learning for relational MDPs. In: Machine Learning Conference of Belgium and the Netherlands (BeNeLearn 2004) (2004)

    Google Scholar 

  12. Morales, E.: Scaling up reinforcement learning with a relational representation. In: Proceedings of the Workshop on Adaptability in Multi-agent Systems at AORC 2003, Sydney (2003)

    Google Scholar 

  13. Guestrin, C., Koller, D., Gearhart, C., Kanodia, N.: Generalizing plans to new environments in relational MDPs. In: IJCAI 2003 (2003)

    Google Scholar 

  14. Kersting, K., De Raedt, L.: Logical markov decision programs and the convergence of logical TD(λ). In: Camacho, R., King, R., Srinivasan, A. (eds.) ILP 2004. LNCS, vol. 3194, pp. 180–197. Springer, Heidelberg (2004)

    Chapter  Google Scholar 

  15. Nienhuys-Cheng, S.-H., de Wolf, R.: Foundations of Inductive Logic Programming. Lecture Notes in Artifical Intelligence, vol. 1228. Springer, Heidelberg (1997)

    Google Scholar 

  16. Clark, K.L.: Negation as failure. In: Logic and Data Bases, pp. 293–322 (1977)

    Google Scholar 

  17. Van Otterlo, M., Kersting, K.: Challenges for relational reinforcement learning. In: ICML 2004 Workshop on Relational Reinforcement Learning (2004)

    Google Scholar 

  18. Slaney, J., Thiébaux, S.: Blocks world revisited. Artificial Intelligence 125, 119–153 (2001)

    Article  MATH  MathSciNet  Google Scholar 

  19. Kersting, K., Van Otterlo, M., De Raedt, L.: Bellman goes relational. In: ICML 2004 (2004)

    Google Scholar 

  20. Lecoeuche, R.: Learning optimal dialogue management rules by using reinforcement learning and inductive logic programming. In: Proc. of the North American Chapter of the Association for Computational Linguistics (NAACL) (2001)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2006 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Zhiwei, S., Xiaoping, C. (2006). Unique State and Automatical Action Abstracting Based on Logical MDPs with Negation. In: Jiao, L., Wang, L., Gao, X., Liu, J., Wu, F. (eds) Advances in Natural Computation. ICNC 2006. Lecture Notes in Computer Science, vol 4222. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11881223_118

Download citation

  • DOI: https://doi.org/10.1007/11881223_118

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-45907-1

  • Online ISBN: 978-3-540-45909-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics