arXiv Analytics

Sign in

arXiv:1810.12282 [cs.LG]AbstractReferencesReviewsResources

Assessing Generalization in Deep Reinforcement Learning

Charles Packer, Katelyn Gao, Jernej Kos, Philipp Krähenbühl, Vladlen Koltun, Dawn Song

Published 2018-10-29Version 1

Deep reinforcement learning (RL) has achieved breakthrough results on many tasks, but has been shown to be sensitive to system changes at test time. As a result, building deep RL agents that generalize has become an active research area. Our aim is to catalyze and streamline community-wide progress on this problem by providing the first benchmark and a common experimental protocol for investigating generalization in RL. Our benchmark contains a diverse set of environments and our evaluation methodology covers both in-distribution and out-of-distribution generalization. To provide a set of baselines for future research, we conduct a systematic evaluation of deep RL algorithms, including those that specifically tackle the problem of generalization.

Related articles: Most relevant | Search more
arXiv:1805.03359 [cs.LG] (Published 2018-05-09)
Reward Estimation for Variance Reduction in Deep Reinforcement Learning
arXiv:2106.13799 [cs.LG] (Published 2021-06-25)
Assessing Generalization of SGD via Disagreement
arXiv:1901.02219 [cs.LG] (Published 2019-01-08)
Uncertainty-Based Out-of-Distribution Detection in Deep Reinforcement Learning