arXiv Analytics

Sign in

arXiv:2410.02895 [math.OC]AbstractReferencesReviewsResources

Approximation Schemes for POMPDs with Continuous Spaces and Their Near Optimality

Ali Devran Kara, Erhan Bayraktar, Serdar Yuksel

Published 2024-10-03Version 1

We study an approximation method for partially observed Markov decision processes (POMDPs) with continuous spaces. Belief MDP reduction, which has been the standard approach to study POMDPs requires rigorous approximation methods for practical applications, due to the state space being lifted to the space of probability measures. Generalizing recent work, in this paper we present rigorous approximation methods via discretizing the observation space and constructing a fully observed finite MDP model using a finite length history of the discrete observations and control actions. We show that the resulting policy is near-optimal under some regularity assumptions on the channel, and under certain controlled filter stability requirements for the hidden state process. Furthermore, by quantizing the measurements, we are able to utilize refined filter stability conditions. We also provide a Q learning algorithm that uses a finite memory of discretized information variables, and prove its convergence to the optimality equation of the finite fully observed MDP constructed using the approximation method.

Related articles: Most relevant | Search more
arXiv:1507.05125 [math.OC] (Published 2015-07-17)
On the Optimality of (s,S) Policies
arXiv:2104.11865 [math.OC] (Published 2021-04-24)
Suboptimal coverings for continuous spaces of control tasks
arXiv:2202.02914 [math.OC] (Published 2022-02-07)
Global convergence and optimality of the heavy ball method for non-convex optimization