arXiv Analytics

Sign in

arXiv:1812.08485 [math.OC]AbstractReferencesReviewsResources

First-order algorithms converge faster than $O(1/k)$ on convex problems

Ching-pei Lee, Stephen J. Wright

Published 2018-12-20Version 1

It has been known for many years that both gradient descent and stochastic coordinate descent achieve a global convergence rate of $O(1/k)$ in the objective value, when applied to a scheme for minimizing a Lipschitz-continuously differentiable, unconstrained convex function. In this work, we improve this rate to $o(1/k)$. We extend the result to proximal gradient and proximal coordinate descent on regularized problems to show similar $o(1/k)$ convergence rates. The result is tight in the sense that an $O(1/k^{1+\epsilon})$ rate is not generally attainable for any $\epsilon>0$, for any of these methods.

Related articles: Most relevant | Search more
arXiv:1312.5799 [math.OC] (Published 2013-12-20, updated 2014-03-01)
Accelerated, Parallel and Proximal Coordinate Descent
arXiv:1711.05812 [math.OC] (Published 2017-11-15)
Global convergence rates of augmented Lagrangian methods for constrained convex programming
arXiv:1608.01713 [math.OC] (Published 2016-08-04)
Global Convergence Rate of Proximal Incremental Aggregated Gradient Methods