Abstract
The theory of Markov Decision Processes (MDPs) provides algorithms for generating an optimal policy. For large domains these algorithms become intractable and approximate solutions become necessary. In this paper we extend previous work on approximate planning in large stochastic domains by using automatically-generated non-uniform abstractions which exploit the structure of the state space. We consider a state space expressed as a cross product of sets, or dimensions. We obtain approximate solutions by varying the level of abstraction, selectively ignoring some of the dimensions in some parts of the state space. We describe a modification of a standard policy generation algorithm for the now non-Markovian decision process, which re-calculates values for nearby states based on a locally uniform abstraction for each state. We present methods to automatically generate an initial abstraction based on the domain structure and to automatically modify the non-uniform abstraction. The changes to the abstraction are based on both the current policy and the likelihood of encountering particular states in the future, thereby taking into account the agent’s changing circumstances.
Preview
Unable to display preview. Download preview PDF.
References
J. Baum and A. E. Nicholson. Dynamic non-uniform abstractions for approximate planning in large structured stochastic domains. Technical Report TR 98/18, Department of Computer Science, Monash Uni., 1998.
R. Bellman. Dynamic Programming. Princeton Uni. Press, 1957.
D. P. Bertsekas. Dynamic Programming. Prentice-Hall. Englewood Cliffs, N.J., 1987.
C. Boutilier. Correlated action effects in decision theoretic regression. In Proc. of UAI, pages 30–37, 1997.
C. Boutilier and R. Dearden. Approximating value trees in structured dynamic programming. In Proc. of 13th Int’l Conf. on Machine Learning, pages 54–62, Bari, Italy, 1996.
C. Boutilier, R. Dearden, and M. Goldszmidt. Exploiting structure in policy construction. In Proc. of IJCAI, pages 1104–1111, Montreal, 1995.
T. Dean, L. P. Kaelbling, J. Kirman, and A. E. Nicholson. Planning under time constraints in stochastic domains. Artificial Intelligence, 76(1–2):35–74, 1995.
R. Dearden and C. Boutilier. Abstraction and approximate decision theoretic planning. Artificial Intelligence, 89(1):219–283, 1997.
M. Drummond and J. Bresina. Anytime synthetic projection: Maximizing the probability of goal satisfaction. In Proc. of AAAI-90, pages 138–144. AAAI, 1990.
R. P. Goldman, D. J. Musliner, K. D. Krebsbach, and M. S. Boddy. Dynamic abstraction planning. In Proc. of AAAI 97, 1997.
R. A. Howard. Dynamic Programming and Markov Processes. MIT Press, Cambridge, Massachusetts, 1960.
J. Kirman. Predicting real-time planner performance by domain characterization. PhD thesis, Brown Uni., 1994.
D. J. Musliner, E. H. Durfee, and K. G. Shin. World modeling for the dynamic construction of real-time plans. Artificial Intelligence, 74:83–127, 1995.
A. E. Nicholson and L. P. Kaelbling. Toward approximate planning in very large stochastic domains. In Proc. of Spring Symposium on Decision Theoretic Planning, 1994.
J. R. Quinlan. C4.5: Programs for Machine Learning, Morgan Kaufmann, San Mateo, California, 1993.
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 1998 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Baum, J., Nicholson, A.E. (1998). Dynamic non-uniform abstractions for approximate planning in large structured stochastic domains. In: Lee, HY., Motoda, H. (eds) PRICAI’98: Topics in Artificial Intelligence. PRICAI 1998. Lecture Notes in Computer Science, vol 1531. Springer, Berlin, Heidelberg . https://doi.org/10.1007/BFb0095303
Download citation
DOI: https://doi.org/10.1007/BFb0095303
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-65271-7
Online ISBN: 978-3-540-49461-4
eBook Packages: Springer Book Archive