Skip to main content

Multivariate Decision Tree Function Approximation for Reinforcement Learning

  • Conference paper
Neural Information Processing. Theory and Algorithms (ICONIP 2010)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 6443))

Included in the following conference series:

Abstract

In reinforcement learning, when dimensionality of the state space increases, making use of state abstraction seems inevitable. Among the methods proposed to solve this problem, decision tree based methods could be useful as they provide automatic state abstraction. But existing methods use univariate, therefore axis-aligned, splits in decision nodes, imposing hyper-rectangular partitioning of the state space. In some applications, multivariate splits can generate smaller and more accurate trees. In this paper, we use oblique decision trees as an instance of multivariate trees to implement state abstraction for reinforcement learning agents. Simulation results on mountain car and puddle world tasks show significant improvement in the average received rewards, average number of steps to finish the task, and size of the trees both in learning and test phases.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Tesauro, G.: Temporal difference learning and td-gammon. Communications of the ACM 38(3), 58–68 (1995)

    Article  Google Scholar 

  2. Boyan, J.A., Moore, A.W.: Generalization in reinforcement learning: Safely approximating the value function. In: Tesauro, G., Touretsky, D.S., Leen, T.K. (eds.) Advances in Neural Information Processing Systems, Cambridge, MA, vol. 7 (1995)

    Google Scholar 

  3. Moore, A.W.: The parti-game algorithm for variable resolution reinforcement learning in multidimensional state-spaces. In: Cowan, J.D., Tesauro, G., Alspector, J. (eds.) Advances in Neural Information Processing Systems, vol. 6, pp. 711–718. Morgan Kaufmann Publishers, Inc., San Francisco (1994)

    Google Scholar 

  4. Chapman, D., Kaelbling, L.P.: Input generalization in delayed reinforcement learning: An algorithm and performance comparisons. In: Proceedings of the Twelfth International Joint Conference on Artificial Intelligence, pp. 726–731 (1991)

    Google Scholar 

  5. McCallum, A.: Reinforcement with Selective Perception and Hidden State. PhD thesis, Computer Science Dept., University of Rochester (1995)

    Google Scholar 

  6. Uther, W.R., Veloso, M.M.: Tree Based Discretization for Continuous State Space Reinforcement Learning. In: Proceedings of the National Conference on Artificial Intelligence, pp. 769–775 (1998)

    Google Scholar 

  7. Asadpour, M.: Behavior Design in Microrobots: Hierarchical Reinforcement Learning under Resource Constraints. PHD thesis, EPFL, Lausanne, Switzerland (2006)

    Google Scholar 

  8. Pyeatt, L.D., Howe, A.E.: Decision tree function approximation in reinforcement learning. In: Proceedings of the Third International Symposium on Adaptive Systems: Evolutionary Computation and Probabilistic Graphical Models, pp. 70–77 (2001)

    Google Scholar 

  9. Brodley, C.E., Utgoff, P.E.: Multivariate decision trees. Machine Learning 19, 45–77 (1995)

    MATH  Google Scholar 

  10. Breiman, L., Friedman, J.H., Olshen, R.A., Stone, C.J.: Classification and regression trees. Technical report, Wadsworth International, Monterey, CA (1984)

    Google Scholar 

  11. Murthy, S.K.: On Growing Better Decision Trees from Data, Ph.D. dissertation, Johns Hopkins Univ., Dept. Comput. Sci., Baltimore, MD (1995)

    Google Scholar 

  12. Heath, D., Kasif, S., Salzberg, S.: Learning Oblique Decision Trees. In: Proceedings of the 13th International Joint Conference on Artificial Intelligence, Chambery, France, pp. 1002–1007. Morgan Kaufmann, San Francisco (1993)

    Google Scholar 

  13. Sutton, R.S., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)

    Google Scholar 

  14. Tanner, B., White, A.: RL-Glue: Language-Independent Software for Reinforcement-Learning Experiments. Journal of Machine Learning Research, 2133–2136 (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Saghezchi, H.B., Asadpour, M. (2010). Multivariate Decision Tree Function Approximation for Reinforcement Learning. In: Wong, K.W., Mendis, B.S.U., Bouzerdoum, A. (eds) Neural Information Processing. Theory and Algorithms. ICONIP 2010. Lecture Notes in Computer Science, vol 6443. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-17537-4_83

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-17537-4_83

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-17536-7

  • Online ISBN: 978-3-642-17537-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics