{ "id": "1906.03574", "version": "v1", "published": "2019-06-09T06:00:50.000Z", "updated": "2019-06-09T06:00:50.000Z", "title": "Transfer Learning by Modeling a Distribution over Policies", "authors": [ "Disha Shrivastava", "Eeshan Gunesh Dhekane", "Riashat Islam" ], "comment": "Accepted at the ICML 2019 workshop on Multi-Task and Lifelong Reinforcement Learning", "categories": [ "cs.LG", "cs.AI", "stat.ML" ], "abstract": "Exploration and adaptation to new tasks in a transfer learning setup is a central challenge in reinforcement learning. In this work, we build on the idea of modeling a distribution over policies in a Bayesian deep reinforcement learning setup to propose a transfer strategy. Recent works have shown to induce diversity in the learned policies by maximizing the entropy of a distribution of policies (Bachman et al., 2018; Garnelo et al., 2018) and thus, we postulate that our proposed approach leads to faster exploration resulting in improved transfer learning. We support our hypothesis by demonstrating favorable experimental results on a variety of settings on fully-observable GridWorld and partially observable MiniGrid (Chevalier-Boisvert et al., 2018) environments.", "revisions": [ { "version": "v1", "updated": "2019-06-09T06:00:50.000Z" } ], "analyses": { "keywords": [ "distribution", "bayesian deep reinforcement learning setup", "transfer strategy", "central challenge", "transfer learning setup" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }