{ "id": "2003.13350", "version": "v1", "published": "2020-03-30T11:33:16.000Z", "updated": "2020-03-30T11:33:16.000Z", "title": "Agent57: Outperforming the Atari Human Benchmark", "authors": [ "Adrià Puigdomènech Badia", "Bilal Piot", "Steven Kapturowski", "Pablo Sprechmann", "Alex Vitvitskyi", "Daniel Guo", "Charles Blundell" ], "categories": [ "cs.LG", "stat.ML" ], "abstract": "Atari games have been a long-standing benchmark in the reinforcement learning (RL) community for the past decade. This benchmark was proposed to test general competency of RL algorithms. Previous work has achieved good average performance by doing outstandingly well on many games of the set, but very poorly in several of the most challenging games. We propose Agent57, the first deep RL agent that outperforms the standard human benchmark on all 57 Atari games. To achieve this result, we train a neural network which parameterizes a family of policies ranging from very exploratory to purely exploitative. We propose an adaptive mechanism to choose which policy to prioritize throughout the training process. Additionally, we utilize a novel parameterization of the architecture that allows for more consistent and stable learning.", "revisions": [ { "version": "v1", "updated": "2020-03-30T11:33:16.000Z" } ], "analyses": { "keywords": [ "atari human benchmark", "atari games", "first deep rl agent", "standard human benchmark", "test general competency" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }