arXiv Analytics

Sign in

arXiv:1712.06585 [math.OC]AbstractReferencesReviewsResources

Third-order Smoothness Helps: Even Faster Stochastic Optimization Algorithms for Finding Local Minima

Yaodong Yu, Pan Xu, Quanquan Gu

Published 2017-12-18Version 1

We propose stochastic optimization algorithms that can find local minima faster than existing algorithms for nonconvex optimization problems, by exploiting the third-order smoothness to escape non-degenerate saddle points more efficiently. More specifically, the proposed algorithm only needs $\tilde{O}(\epsilon^{-10/3})$ stochastic gradient evaluations to converge to an approximate local minimum $\mathbf{x}$, which satisfies $\|\nabla f(\mathbf{x})\|_2\leq\epsilon$ and $\lambda_{\min}(\nabla^2 f(\mathbf{x}))\geq -\sqrt{\epsilon}$ in the general stochastic optimization setting, where $\tilde{O}(\cdot)$ hides logarithm polynomial terms and constants. This improves upon the $\tilde{O}(\epsilon^{-7/2})$ gradient complexity achieved by the state-of-the-art stochastic local minima finding algorithms by a factor of $\tilde{O}(\epsilon^{-1/6})$. For nonconvex finite-sum optimization, our algorithm also outperforms the best known algorithms in a certain regime.

Related articles: Most relevant | Search more
arXiv:2008.06148 [math.OC] (Published 2020-08-14)
Complexity aspects of local minima and related notions
arXiv:1611.01146 [math.OC] (Published 2016-11-03)
Finding Local Minima for Nonconvex Optimization in Linear Time
arXiv:2303.03536 [math.OC] (Published 2023-03-06, updated 2023-07-13)
Certifying the absence of spurious local minima at infinity