arXiv Analytics

Sign in

arXiv:1107.1744 [math.OC]AbstractReferencesReviewsResources

Stochastic convex optimization with bandit feedback

Alekh Agarwal, Dean P. Foster, Daniel Hsu, Sham M. Kakade, Alexander Rakhlin

Published 2011-07-08, updated 2011-10-08Version 2

This paper addresses the problem of minimizing a convex, Lipschitz function $f$ over a convex, compact set $\xset$ under a stochastic bandit feedback model. In this model, the algorithm is allowed to observe noisy realizations of the function value $f(x)$ at any query point $x \in \xset$. The quantity of interest is the regret of the algorithm, which is the sum of the function values at algorithm's query points minus the optimal function value. We demonstrate a generalization of the ellipsoid algorithm that incurs $\otil(\poly(d)\sqrt{T})$ regret. Since any algorithm has regret at least $\Omega(\sqrt{T})$ on this problem, our algorithm is optimal in terms of the scaling with $T$.

Related articles: Most relevant | Search more
arXiv:2107.00469 [math.OC] (Published 2021-06-29)
Never Go Full Batch (in Stochastic Convex Optimization)
arXiv:2402.03210 [math.OC] (Published 2024-02-05, updated 2024-07-11)
Universal Gradient Methods for Stochastic Convex Optimization
arXiv:2207.02750 [math.OC] (Published 2022-07-06)
An SDE perspective on stochastic convex optimization