Skip to main content

Dynamic non-uniform abstractions for approximate planning in large structured stochastic domains

  • Stochastic Process
  • Conference paper
  • First Online:
PRICAI’98: Topics in Artificial Intelligence (PRICAI 1998)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1531))

Included in the following conference series:

  • 98 Accesses

Abstract

The theory of Markov Decision Processes (MDPs) provides algorithms for generating an optimal policy. For large domains these algorithms become intractable and approximate solutions become necessary. In this paper we extend previous work on approximate planning in large stochastic domains by using automatically-generated non-uniform abstractions which exploit the structure of the state space. We consider a state space expressed as a cross product of sets, or dimensions. We obtain approximate solutions by varying the level of abstraction, selectively ignoring some of the dimensions in some parts of the state space. We describe a modification of a standard policy generation algorithm for the now non-Markovian decision process, which re-calculates values for nearby states based on a locally uniform abstraction for each state. We present methods to automatically generate an initial abstraction based on the domain structure and to automatically modify the non-uniform abstraction. The changes to the abstraction are based on both the current policy and the likelihood of encountering particular states in the future, thereby taking into account the agent’s changing circumstances.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. J. Baum and A. E. Nicholson. Dynamic non-uniform abstractions for approximate planning in large structured stochastic domains. Technical Report TR 98/18, Department of Computer Science, Monash Uni., 1998.

    Google Scholar 

  2. R. Bellman. Dynamic Programming. Princeton Uni. Press, 1957.

    Google Scholar 

  3. D. P. Bertsekas. Dynamic Programming. Prentice-Hall. Englewood Cliffs, N.J., 1987.

    Google Scholar 

  4. C. Boutilier. Correlated action effects in decision theoretic regression. In Proc. of UAI, pages 30–37, 1997.

    Google Scholar 

  5. C. Boutilier and R. Dearden. Approximating value trees in structured dynamic programming. In Proc. of 13th Int’l Conf. on Machine Learning, pages 54–62, Bari, Italy, 1996.

    Google Scholar 

  6. C. Boutilier, R. Dearden, and M. Goldszmidt. Exploiting structure in policy construction. In Proc. of IJCAI, pages 1104–1111, Montreal, 1995.

    Google Scholar 

  7. T. Dean, L. P. Kaelbling, J. Kirman, and A. E. Nicholson. Planning under time constraints in stochastic domains. Artificial Intelligence, 76(1–2):35–74, 1995.

    Article  Google Scholar 

  8. R. Dearden and C. Boutilier. Abstraction and approximate decision theoretic planning. Artificial Intelligence, 89(1):219–283, 1997.

    Article  MATH  MathSciNet  Google Scholar 

  9. M. Drummond and J. Bresina. Anytime synthetic projection: Maximizing the probability of goal satisfaction. In Proc. of AAAI-90, pages 138–144. AAAI, 1990.

    Google Scholar 

  10. R. P. Goldman, D. J. Musliner, K. D. Krebsbach, and M. S. Boddy. Dynamic abstraction planning. In Proc. of AAAI 97, 1997.

    Google Scholar 

  11. R. A. Howard. Dynamic Programming and Markov Processes. MIT Press, Cambridge, Massachusetts, 1960.

    MATH  Google Scholar 

  12. J. Kirman. Predicting real-time planner performance by domain characterization. PhD thesis, Brown Uni., 1994.

    Google Scholar 

  13. D. J. Musliner, E. H. Durfee, and K. G. Shin. World modeling for the dynamic construction of real-time plans. Artificial Intelligence, 74:83–127, 1995.

    Article  Google Scholar 

  14. A. E. Nicholson and L. P. Kaelbling. Toward approximate planning in very large stochastic domains. In Proc. of Spring Symposium on Decision Theoretic Planning, 1994.

    Google Scholar 

  15. J. R. Quinlan. C4.5: Programs for Machine Learning, Morgan Kaufmann, San Mateo, California, 1993.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Hing-Yan Lee Hiroshi Motoda

Rights and permissions

Reprints and permissions

Copyright information

© 1998 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Baum, J., Nicholson, A.E. (1998). Dynamic non-uniform abstractions for approximate planning in large structured stochastic domains. In: Lee, HY., Motoda, H. (eds) PRICAI’98: Topics in Artificial Intelligence. PRICAI 1998. Lecture Notes in Computer Science, vol 1531. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0095303

Download citation

  • DOI: https://doi.org/10.1007/BFb0095303

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-65271-7

  • Online ISBN: 978-3-540-49461-4

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics