arXiv Analytics

Sign in

arXiv:2211.15457 [cs.LG]AbstractReferencesReviewsResources

Hypernetworks for Zero-shot Transfer in Reinforcement Learning

Sahand Rezaei-Shoshtari, Charlotte Morissette, Francois Robert Hogan, Gregory Dudek, David Meger

Published 2022-11-28, updated 2023-01-02Version 2

In this paper, hypernetworks are trained to generate behaviors across a range of unseen task conditions, via a novel TD-based training objective and data from a set of near-optimal RL solutions for training tasks. This work relates to meta RL, contextual RL, and transfer learning, with a particular focus on zero-shot performance at test time, enabled by knowledge of the task parameters (also known as context). Our technical approach is based upon viewing each RL algorithm as a mapping from the MDP specifics to the near-optimal value function and policy and seek to approximate it with a hypernetwork that can generate near-optimal value functions and policies, given the parameters of the MDP. We show that, under certain conditions, this mapping can be considered as a supervised learning problem. We empirically evaluate the effectiveness of our method for zero-shot transfer to new reward and transition dynamics on a series of continuous control tasks from DeepMind Control Suite. Our method demonstrates significant improvements over baselines from multitask and meta RL approaches.

Related articles: Most relevant | Search more
arXiv:1803.00590 [cs.LG] (Published 2018-03-01)
Hierarchical Imitation and Reinforcement Learning
arXiv:1609.09106 [cs.LG] (Published 2016-09-27)
HyperNetworks
arXiv:1809.01560 [cs.LG] (Published 2018-09-05)
Reinforcement Learning under Threats