arXiv Analytics

Sign in

arXiv:2212.06355 [stat.ML]AbstractReferencesReviewsResources

A Review of Off-Policy Evaluation in Reinforcement Learning

Masatoshi Uehara, Chengchun Shi, Nathan Kallus

Published 2022-12-13Version 1

Reinforcement learning (RL) is one of the most vibrant research frontiers in machine learning and has been recently applied to solve a number of challenging problems. In this paper, we primarily focus on off-policy evaluation (OPE), one of the most fundamental topics in RL. In recent years, a number of OPE methods have been developed in the statistics and computer science literature. We provide a discussion on the efficiency bound of OPE, some of the existing state-of-the-art OPE methods, their statistical properties and some other related research directions that are currently actively explored.

Related articles: Most relevant | Search more
arXiv:2306.04836 [stat.ML] (Published 2023-06-07)
$K$-Nearest-Neighbor Resampling for Off-Policy Evaluation in Stochastic Control
arXiv:2006.06982 [stat.ML] (Published 2020-06-12)
Confidence Interval for Off-Policy Evaluation from Dependent Samples via Bandit Algorithm: Approach from Standardized Martingales
arXiv:2108.04763 [stat.ML] (Published 2021-08-10)
Imitation Learning by Reinforcement Learning