{ "id": "1810.09103", "version": "v1", "published": "2018-10-22T06:35:03.000Z", "updated": "2018-10-22T06:35:03.000Z", "title": "Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces", "authors": [ "Sungsu Lim", "Ajin Joseph", "Lei Le", "Yangchen Pan", "Martha White" ], "categories": [ "cs.LG", "cs.AI", "stat.ML" ], "abstract": "Value-based approaches can be difficult to use in continuous action spaces, because an optimization has to be solved to find the greedy action for the action-values. A common strategy has been to restrict the functional form of the action-values to be convex or quadratic in the actions, to simplify this optimization. Such restrictions, however, can prevent learning accurate action-values. In this work, we propose the Actor-Expert framework for value-based methods, that decouples action-selection (Actor) from the action-value representation (Expert). The Expert uses Q-learning to update the action-values towards the optimal action-values, whereas the Actor (learns to) output the greedy action for the current action-values. We develop a Conditional Cross Entropy Method for the Actor, to learn the greedy action for a generically parameterized Expert, and provide a two-timescale analysis to validate asymptotic behavior. We demonstrate in a toy domain with bimodal action-values that previous restrictive action-value methods fail whereas the decoupled Actor-Expert with a more general action-value parameterization succeeds. Finally, we demonstrate that Actor-Expert performs as well as or better than these other methods on several benchmark continuous-action domains.", "revisions": [ { "version": "v1", "updated": "2018-10-22T06:35:03.000Z" } ], "analyses": { "keywords": [ "continuous action spaces", "actor-expert", "greedy action", "general action-value parameterization succeeds", "conditional cross entropy method" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }