arXiv Analytics

Sign in

arXiv:2009.04258 [math.OC]AbstractReferencesReviewsResources

Bandit Online Learning of Nash Equilibria in Monotone Games

Tatiana Tatarenko, Maryam Kamgarpour

Published 2020-09-08Version 1

We address online bandit learning of Nash equilibria in multi-agent convex games. We propose an algorithm whereby each agent uses only obtained values of her cost function at each joint played action, lacking any information of the functional form of her cost or other agents' costs or strategies. In contrast to past work where convergent algorithms required strong monotonicity, we prove that the algorithm converges to a Nash equilibrium under mere monotonicity assumption. The proposed algorithm extends the applicability of bandit learning in several games including zero-sum convex games with possibly unbounded action spaces, mixed extension of finite-action zero-sum games, as well as convex games with linear coupling constraints.

Comments: arXiv admin note: text overlap with arXiv:1904.01882
Categories: math.OC
Related articles: Most relevant | Search more
arXiv:1802.08412 [math.OC] (Published 2018-02-23)
A Game Problem for Heat Equation
arXiv:2104.11096 [math.OC] (Published 2021-04-22)
On the exact convergence to Nash equilibrium in hypomonotone regimes under full and partial-information
arXiv:1505.01328 [math.OC] (Published 2015-05-06)
An $ε$-Nash equilibrium with high probability for strategic customers in heavy traffic