arXiv Analytics

Sign in

arXiv:2101.09741 [math.OC]AbstractReferencesReviewsResources

An optimal gradient method for smooth (possibly strongly) convex minimization

Adrien Taylor, Yoel Drori

Published 2021-01-24Version 1

We present an optimal gradient method for smooth (possibly strongly) convex optimization. The method is optimal in the sense that its worst-case bound exactly matches the lower bound on the oracle complexity for the class of problems, meaning that no black-box first-order method can have a better worst-case guarantee without further assumptions on the class of problems at hand. The method is in some sense a generalization of the Optimized Gradient Method of Kim and Fessler (2016), and asymptotically corresponds to the Triple Momentum Method (Van Scoy et al., 2017), in the presence of strong convexity. Furthermore, the method is numerically stable to arbitrarily large condition numbers and admits a conceptually very simple proof, which involves a Lyapunov argument and a sum of two inequalities. Finally, we provide a numerical recipe for obtaining the algorithmic parameters of the method, using semidefinite programming, and illustrate that it can be used for developing other methods as well.

Comments: Codes available at https://github.com/AdrienTaylor/Optimal-Gradient-Method (symbolic verifications and numerical experiments)
Categories: math.OC, cs.NA, math.NA
Related articles: Most relevant | Search more
arXiv:1303.4645 [math.OC] (Published 2013-03-19, updated 2013-09-08)
Gradient methods for convex minimization: better rates under weaker conditions
arXiv:1803.02919 [math.OC] (Published 2018-03-07)
Proximal Activation of Smooth Functions in Splitting Algorithms for Convex Minimization
arXiv:1609.09441 [math.OC] (Published 2016-09-29)
Fast dual proximal gradient algorithms with rate $O(1/k^{1.5})$ for convex minimization