{ "id": "2009.04258", "version": "v1", "published": "2020-09-08T13:05:25.000Z", "updated": "2020-09-08T13:05:25.000Z", "title": "Bandit Online Learning of Nash Equilibria in Monotone Games", "authors": [ "Tatiana Tatarenko", "Maryam Kamgarpour" ], "comment": "arXiv admin note: text overlap with arXiv:1904.01882", "categories": [ "math.OC" ], "abstract": "We address online bandit learning of Nash equilibria in multi-agent convex games. We propose an algorithm whereby each agent uses only obtained values of her cost function at each joint played action, lacking any information of the functional form of her cost or other agents' costs or strategies. In contrast to past work where convergent algorithms required strong monotonicity, we prove that the algorithm converges to a Nash equilibrium under mere monotonicity assumption. The proposed algorithm extends the applicability of bandit learning in several games including zero-sum convex games with possibly unbounded action spaces, mixed extension of finite-action zero-sum games, as well as convex games with linear coupling constraints.", "revisions": [ { "version": "v1", "updated": "2020-09-08T13:05:25.000Z" } ], "analyses": { "keywords": [ "nash equilibrium", "bandit online learning", "monotone games", "finite-action zero-sum games", "address online bandit" ], "note": { "typesetting": "TeX", "pages": 0, "language": "en", "license": "arXiv", "status": "editable" } } }