arXiv Analytics

Sign in

arXiv:2201.07908 [math.OC]AbstractReferencesReviewsResources

Markov decision processes with observation costs

Christoph Reisinger, Jonathan Tam

Published 2022-01-19Version 1

We present a framework for a controlled Markov chain where the state of the chain is only given at chosen observation times and of a cost. Optimal strategies therefore involve the choice of observation times as well as the subsequent control values. We show that the corresponding value function satisfies a dynamic programming principle, which leads to a system of quasi-variational inequalities (QVIs). Next, we give an extension where the model parameters are not known a priori but are inferred from the costly observations by Bayesian updates. We then prove a comparison principle for a larger class of QVIs, which implies uniqueness of solutions to our proposed problem. We utilise penalty methods to obtain arbitrarily accurate solutions. Finally, we perform numerical experiments on three applications which illustrate our framework.

Related articles: Most relevant | Search more
arXiv:1904.05481 [math.OC] (Published 2019-04-10)
Stochastic Comparative Statics in Markov Decision Processes
arXiv:1202.6259 [math.OC] (Published 2012-02-28)
A distance for probability spaces, and long-term values in Markov Decision Processes and Repeated Games
arXiv:1310.7906 [math.OC] (Published 2013-10-29, updated 2015-08-04)
Convergence Analysis of the Approximate Newton Method for Markov Decision Processes