arXiv Analytics

Sign in

arXiv:1905.08513 [cs.LG]AbstractReferencesReviewsResources

Stochastic Inverse Reinforcement Learning

Ce Ju, Dong Eui Chang

Published 2019-05-21Version 1

Inverse reinforcement learning (IRL) is an ill-posed inverse problem since expert demonstrations may infer many solutions of reward functions which is hard to recover by local search methods such as a gradient method. In this paper, we generalize the original IRL problem to recover a probability distribution for reward functions. We call such a generalized problem stochastic inverse reinforcement learning (SIRL) which is first formulated as an expectation optimization problem. We adopt the Monte Carlo expectation-maximization (MCEM) method, a global search method, to estimate the parameter of the probability distribution as the first solution to SIRL. With our approach, it is possible to observe the deep intrinsic property in IRL from a global viewpoint, and the technique achieves a considerable robust recovery performance on the classic learning environment, objectworld.

Related articles: Most relevant | Search more
arXiv:1708.01611 [cs.LG] (Published 2017-08-04)
Identification of Probabilities
arXiv:2003.03778 [cs.LG] (Published 2020-03-08)
Adversarial Attacks on Probabilistic Autoregressive Forecasting Models
arXiv:2102.04593 [cs.LG] (Published 2021-02-09)
Regularized Generative Adversarial Network