arXiv Analytics

Sign in

arXiv:1901.10682 [math.OC]AbstractReferencesReviewsResources

On the Convergence of (Stochastic) Gradient Descent with Extrapolation for Non-Convex Optimization

Yi Xu, Zhuoning Yuan, Sen Yang, Rong Jin, Tianbao Yang

Published 2019-01-30Version 1

Extrapolation is very popular in convex optimization, and even for non-convex optimizaion, several recent works have empirically shown its success in many machine learning tasks. However, it has not been analyzed for non-convex optimization and there still remains a gap between the theory and the practice. In this paper, we analyze gradient descent with extrapolation for non-convex optimization both in deterministic and stochastic settings. To the best of our knowledge, this is the first attempt to analyze GD with extrapolation both for non-convex deterministic and stochastic optimization.

Related articles: Most relevant | Search more
arXiv:1310.7063 [math.OC] (Published 2013-10-26, updated 2015-07-01)
On the Convergence of Decentralized Gradient Descent
arXiv:0803.2211 [math.OC] (Published 2008-03-14, updated 2010-05-09)
On Conditions for Convergence to Consensus
arXiv:1801.08691 [math.OC] (Published 2018-01-26)
On Quasi-Newton Forward--Backward Splitting: Proximal Calculus and Convergence