Skip to main content
Log in

Asymptotic linear programming and policy improvement for singularly perturbed Markov decision processes

  • Published:
Mathematical Methods of Operations Research Aims and scope Submit manuscript

Abstract.

In this paper we consider a singularly perturbed Markov decision process with finitely many states and actions and the limiting expected average reward criterion. We make no assumptions about the underlying ergodic structure. We present algorithms for the computation of a uniformly optimal deterministic control, that is, a control which is optimal for all values of the perturbation parameter that are sufficiently small. Our algorithms are based on Jeroslow's Asymptotic Linear Programming.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Author information

Authors and Affiliations

Authors

Additional information

Manuscript received: July 1997/final version received: July 1998

Rights and permissions

Reprints and permissions

About this article

Cite this article

Altman, E., Avrachenkov, K. & Filar, J. Asymptotic linear programming and policy improvement for singularly perturbed Markov decision processes. Mathematical Methods of OR 49, 97–109 (1999). https://doi.org/10.1007/s001860050015

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1007/s001860050015

Navigation