arXiv Analytics

Sign in

arXiv:1806.01175 [cs.LG]AbstractReferencesReviewsResources

TD or not TD: Analyzing the Role of Temporal Differencing in Deep Reinforcement Learning

Artemij Amiranashvili, Alexey Dosovitskiy, Vladlen Koltun, Thomas Brox

Published 2018-06-04Version 1

Our understanding of reinforcement learning (RL) has been shaped by theoretical and empirical results that were obtained decades ago using tabular representations and linear function approximators. These results suggest that RL methods that use temporal differencing (TD) are superior to direct Monte Carlo estimation (MC). How do these results hold up in deep RL, which deals with perceptually complex environments and deep nonlinear models? In this paper, we re-examine the role of TD in modern deep RL, using specially designed environments that control for specific factors that affect performance, such as reward sparsity, reward delay, and the perceptual complexity of the task. When comparing TD with infinite-horizon MC, we are able to reproduce classic results in modern settings. Yet we also find that finite-horizon MC is not inferior to TD, even when rewards are sparse or delayed. This makes MC a viable alternative to TD in deep RL.

Related articles: Most relevant | Search more
arXiv:1707.00524 [cs.LG] (Published 2017-07-03)
Hashing Over Predicted Future Frames for Informed Exploration of Deep Reinforcement Learning
arXiv:1805.11088 [cs.LG] (Published 2018-05-26)
Deep Reinforcement Learning in Ice Hockey for Context-Aware Player Evaluation
arXiv:1901.02219 [cs.LG] (Published 2019-01-08)
Uncertainty-Based Out-of-Distribution Detection in Deep Reinforcement Learning