arXiv Analytics

Sign in

arXiv:2006.11144 [math.OC]AbstractReferencesReviewsResources

On the Almost Sure Convergence of Stochastic Gradient Descent in Non-Convex Problems

Panayotis Mertikopoulos, Nadav Hallak, Ali Kavis, Volkan Cevher

Published 2020-06-19Version 1

This paper analyzes the trajectories of stochastic gradient descent (SGD) to help understand the algorithm's convergence properties in non-convex problems. We first show that the sequence of iterates generated by SGD remains bounded and converges with probability $1$ under a very broad range of step-size schedules. Subsequently, going beyond existing positive probability guarantees, we show that SGD avoids strict saddle points/manifolds with probability $1$ for the entire spectrum of step-size policies considered. Finally, we prove that the algorithm's rate of convergence to Hurwicz minimizers is $\mathcal{O}(1/n^{p})$ if the method is employed with a $\Theta(1/n^p)$ step-size schedule. This provides an important guideline for tuning the algorithm's step-size as it suggests that a cool-down phase with a vanishing step-size could lead to faster convergence; we demonstrate this heuristic using ResNet architectures on CIFAR.

Related articles: Most relevant | Search more
arXiv:2406.09241 [math.OC] (Published 2024-06-13)
What is the long-run distribution of stochastic gradient descent? A large deviations analysis
arXiv:2412.06070 [math.OC] (Published 2024-12-08)
Stochastic Gradient Descent Revisited
arXiv:1709.04718 [math.OC] (Published 2017-09-14)
The Impact of Local Geometry and Batch Size on the Convergence and Divergence of Stochastic Gradient Descent