arXiv Analytics

Sign in

arXiv:1910.12453 [cs.LG]AbstractReferencesReviewsResources

Asynchronous Methods for Model-Based Reinforcement Learning

Yunzhi Zhang, Ignasi Clavera, Boren Tsai, Pieter Abbeel

Published 2019-10-28Version 1

Significant progress has been made in the area of model-based reinforcement learning. State-of-the-art algorithms are now able to match the asymptotic performance of model-free methods while being significantly more data efficient. However, this success has come at a price: state-of-the-art model-based methods require significant computation interleaved with data collection, resulting in run times that take days, even if the amount of agent interaction might be just hours or even minutes. When considering the goal of learning in real-time on real robots, this means these state-of-the-art model-based algorithms still remain impractical. In this work, we propose an asynchronous framework for model-based reinforcement learning methods that brings down the run time of these algorithms to be just the data collection time. We evaluate our asynchronous framework on a range of standard MuJoCo benchmarks. We also evaluate our asynchronous framework on three real-world robotic manipulation tasks. We show how asynchronous learning not only speeds up learning w.r.t wall-clock time through parallelization, but also further reduces the sample complexity of model-based approaches by means of improving the exploration and by means of effectively avoiding the policy overfitting to the deficiencies of learned dynamics models.

Related articles: Most relevant | Search more
arXiv:2004.08648 [cs.LG] (Published 2020-04-18)
Modeling Survival in model-based Reinforcement Learning
arXiv:2002.04523 [cs.LG] (Published 2020-02-11)
Objective Mismatch in Model-based Reinforcement Learning
arXiv:1807.03858 [cs.LG] (Published 2018-07-10)
Algorithmic Framework for Model-based Reinforcement Learning with Theoretical Guarantees