arXiv Analytics

Sign in

arXiv:1903.05812 [math.OC]AbstractReferencesReviewsResources

Reinforcement Learning for Decentralized Stochastic Control and Coordination Games

Bora Yongacoglu, Gürdal Arslan, Serdar Yüksel

Published 2019-03-14Version 1

In the study of stochastic dynamic team problems, analytical methods for finding optimal policies are often inapplicable due to lack of prior knowledge of the cost function or the state dynamics. Reinforcement learning offers a possible solution to such coordination problems. Existing learning methods for coordinating play either rely on control sharing among controllers or otherwise, in general, do not guarantee convergence to optimal policies. In a recent paper, we provided a decentralized algorithm for finding equilibrium policies in weakly acyclic stochastic dynamic games, which contain team games as an important special case. However, stochastic dynamic teams can in general possess suboptimal equilibrium policies whose cost can be arbitrarily higher than a team optimal policy's cost. In this paper, we present a reinforcement learning algorithm and its refinements, and provide probabilistic guarantees for convergence to globally optimal policies in team games as well as a more general class of coordination games. The algorithms presented here are strictly decentralized in that they require only access to local information such as cost realizations, previous local actions, and state transitions.

Related articles: Most relevant | Search more
arXiv:1802.07668 [math.OC] (Published 2018-02-21)
A model for system uncertainty in reinforcement learning
arXiv:1803.05811 [math.OC] (Published 2018-03-15)
On a General Dynamic Programming Approach for Decentralized Stochastic Control
arXiv:1809.08712 [math.OC] (Published 2018-09-24)
Decentralized Stochastic Control with a Word-of-Mouth Communication