arXiv Analytics

Sign in

arXiv:2009.08586 [cs.LG]AbstractReferencesReviewsResources

A Contraction Approach to Model-based Reinforcement Learning

Ting-Han Fan, Peter J. Ramadge

Published 2020-09-18Version 1

Model-based Reinforcement Learning has shown considerable experimental success. However, a theoretical understanding of it is still lacking. To this end, we analyze the error in cumulative reward for both stochastic and deterministic transitions using a contraction approach. We show that this approach doesn't require strong assumptions and can recover the typical quadratic error to the horizon. We prove that branched rollouts can reduce this error and are essential for deterministic transitions to have a Bellman contraction. Our results also apply to Imitation Learning, where we prove that GAN-type learning is better than Behavioral Cloning in continuous state and action spaces.

Related articles: Most relevant | Search more
arXiv:2106.14080 [cs.LG] (Published 2021-06-26)
Model-Advantage Optimization for Model-Based Reinforcement Learning
arXiv:2006.16712 [cs.LG] (Published 2020-06-30)
Model-based Reinforcement Learning: A Survey
arXiv:2411.11511 [cs.LG] (Published 2024-11-18)
Structure learning with Temporal Gaussian Mixture for model-based Reinforcement Learning