arXiv Analytics

Sign in

arXiv:2107.00469 [math.OC]AbstractReferencesReviewsResources

Never Go Full Batch (in Stochastic Convex Optimization)

Idan Amir, Yair Carmon, Tomer Koren, Roi Livni

Published 2021-06-29Version 1

We study the generalization performance of $\text{full-batch}$ optimization algorithms for stochastic convex optimization: these are first-order methods that only access the exact gradient of the empirical risk (rather than gradients with respect to individual data points), that include a wide range of algorithms such as gradient descent, mirror descent, and their regularized and/or accelerated variants. We provide a new separation result showing that, while algorithms such as stochastic gradient descent can generalize and optimize the population risk to within $\epsilon$ after $O(1/\epsilon^2)$ iterations, full-batch methods either need at least $\Omega(1/\epsilon^4)$ iterations or exhibit a dimension-dependent sample complexity.

Related articles: Most relevant | Search more
arXiv:2106.07644 [math.OC] (Published 2021-06-10)
A Continuized View on Nesterov Acceleration for Stochastic Gradient Descent and Randomized Gossip
Mathieu Even et al.
arXiv:2505.11343 [math.OC] (Published 2025-05-16)
Revisiting Stochastic Approximation and Stochastic Gradient Descent
arXiv:2406.09241 [math.OC] (Published 2024-06-13)
What is the long-run distribution of stochastic gradient descent? A large deviations analysis