arXiv Analytics

Sign in

arXiv:1804.07193 [cs.LG]AbstractReferencesReviewsResources

Lipschitz Continuity in Model-based Reinforcement Learning

Kavosh Asadi, Dipendra Misra, Michael L. Littman

Published 2018-04-19Version 1

Model-based reinforcement-learning methods learn transition and reward models and use them to guide behavior. We analyze the impact of learning models that are Lipschitz continuous---the distance between function values for two inputs is bounded by a linear function of the distance between the inputs. Our first result shows a tight bound on model errors for multi-step predictions with Lipschitz continuous models. We go on to prove an error bound for the value-function estimate arising from such models and show that the estimated value function is itself Lipschitz continuous. We conclude with empirical results that demonstrate significant benefits to enforcing Lipschitz continuity of neural net models during reinforcement learning.

Related articles: Most relevant | Search more
arXiv:2106.14080 [cs.LG] (Published 2021-06-26)
Model-Advantage Optimization for Model-Based Reinforcement Learning
arXiv:1904.10762 [cs.LG] (Published 2019-04-23)
Baconian: A Unified Opensource Framework for Model-Based Reinforcement Learning
arXiv:1909.11821 [cs.LG] (Published 2019-09-25)
Model Imitation for Model-Based Reinforcement Learning