arXiv Analytics

Sign in

arXiv:1809.02064 [cs.LG]AbstractReferencesReviewsResources

Sample-Efficient Imitation Learning via Generative Adversarial Nets

Lionel Blondé, Alexandros Kalousis

Published 2018-09-06Version 1

Recent work in imitation learning articulate their formulation around the GAIL architecture, relying on the adversarial training procedure introduced in GANs. Albeit successful at generating behaviours similar to those demonstrated to the agent, GAIL suffers from a high sample complexity in the number of interactions it has to carry out in the environment in order to achieve satisfactory performance. In this work, we dramatically shrink the amount of interactions with the environment by leveraging an off-policy actor-critic architecture. Additionally, employing deterministic policy gradients allows us to treat the learned reward as a differentiable node in the computational graph, while preserving the model-free nature of our approach. Our experiments span a variety of continuous control tasks.

Related articles: Most relevant | Search more
arXiv:1711.08267 [cs.LG] (Published 2017-11-22)
GraphGAN: Graph Representation Learning with Generative Adversarial Nets
Hongwei Wang et al.
arXiv:1806.02920 [cs.LG] (Published 2018-06-07)
GAIN: Missing Data Imputation using Generative Adversarial Nets
arXiv:1705.08395 [cs.LG] (Published 2017-05-23)
Continual Learning in Generative Adversarial Nets