Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Markov decision processes under observability constraints

  • Published:
Mathematical Methods of Operations Research Aims and scope Submit manuscript

Abstract

We develop an algorithm to compute optimal policies for Markov decision processes subject to constraints that result from some observability restrictions on the process. We assume that the state of the Markov process is unobservable. There is an observable process related to the unobservable state. So, we want to find a decision rule depending only on this observable process. The objective is to minimize the expected average cost over an infinite horizon. We also analyze the possibility of performing observations in more detail to obtain improved policies.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yasemin Serin.

Additional information

Manuscript received: March 2004/Final version received: June 2004

Rights and permissions

Reprints and permissions

About this article

Cite this article

Serin, Y., Kulkarni, V. Markov decision processes under observability constraints. Math Meth Oper Res 61, 311–328 (2005). https://doi.org/10.1007/s001860400402

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1007/s001860400402

Keywords