Abstract
We give an example from the theory of Markov decision processes which shows that the “optimism in the face of uncertainty” heuristics may fail to make any progress. This is due to the impossibility to falsify a belief that a (transition) probability is larger than 0. Our example shows the utility of Popper’s demand of falsifiability of hypotheses in the area of artificial intelligence.



Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
The transition matrix P is the matrix of transition probabilities with rows and columns indexed by the states in S, so that in row s and column s′ the entry in P is the transition probability p(s, s′) from s to s′.
Note that the claim that the outcome of a random experiment has positive probability is basically an existence claim over an infinite number of trials (and hence not refutable).
Incidentally, this can also be used as criticism on Pascal’s wager. Although it is not clear whether it is appropriate to represent Pascal’s wager as an MDP similar to that in Example 1, Pascal’s argument is based on a non-refutable belief, as the assumption that there is a positive transition probability to heaven is not falsifiable.
That is, for an optimistic estimate of the transition probability p in question, one computes the expected reward which may be gained when insisting in the transition. This can be compared to the reward to be expected when ignoring the transition (i.e. setting p = 0 in the agent’s model). If the latter value is larger, the agent refutes the hypothesis that p > 0.
It is worth noting that although the UCRL algorithm assumes that the underlying MDP is ergodic or communicating, the optimistic model of the MDP it assumes in general is neither ergodic nor communicating.
References
Auer, P., & Ortner, R. (2006). Logarithmic online regret bounds for reinforcement learning. In B. Schölkopf, J. C. Platt, & T. Hofmann (Eds.), Advances in Neural Information Processing Systems (Vol. 19, pp. 49–56). Cambridge, MA: MIT Press.
Auer, P., Jaksch, T., & Ortner, R. (2008). Near-optimal regret bounds for reinforcement learning (submitted).
Brafman, R. I., & Tennenholtz, M. (2002). R-max – a general polynomial time algorithm for near-optimal reinforcement learning. Journal of Machine Learning Research, 3, 213–231.
Kemeny, J. G., Snell, J. L., & Knapp, A. W. (1976). Denumerable Markov chains. New York: Springer.
Popper, K. R. (1969). Logik der Forschung (3rd ed.). Tübingen: Mohr.
Puterman, M. L. (1994). Markov decision processes. Discrete stochastic programming. New York: Wiley.
Strehl, A. L., & Littman, M. L. (2004). An empirical evaluation of interval estimation for Markov decision processes. In 16th IEEE International Conference on Tools with Artificial Intelligence (ICTAI 2004) (pp.128–135). IEEE Computer Society.
Strehl, A. L., & Littman, M. L. (2005). A theoretical analysis of model-based interval estimation. In L. De Raedt & S.Wrobel (Eds.), Machine learning, Proceedings of the Twenty-Second International Conference (ICML 2005) (pp. 857–864) ACM.
Acknowledgments
The author would like to thank Georg Dorn for comments on a prior version of this paper. This work was supported in part by the Austrian Science Fund FWF (S9104-N13 SP4) and the IST Programme of the European Community, under the PASCAL Network of Excellence, IST-2002-506778. This publication only reflects the authors’ views.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Ortner, R. Optimism in the Face of Uncertainty Should be Refutable. Minds & Machines 18, 521–526 (2008). https://doi.org/10.1007/s11023-008-9115-5
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11023-008-9115-5