arXiv Analytics

Sign in

arXiv:2402.06271 [math.OC]AbstractReferencesReviewsResources

Adaptive proximal gradient methods are universal without approximation

Konstantinos A. Oikonomidis, Emanuel Laude, Puya Latafat, Andreas Themelis, Panagiotis Patrinos

Published 2024-02-09, updated 2024-07-05Version 2

We show that adaptive proximal gradient methods for convex problems are not restricted to traditional Lipschitzian assumptions. Our analysis reveals that a class of linesearch-free methods is still convergent under mere local H\"older gradient continuity, covering in particular continuously differentiable semi-algebraic functions. To mitigate the lack of local Lipschitz continuity, popular approaches revolve around $\varepsilon$-oracles and/or linesearch procedures. In contrast, we exploit plain H\"older inequalities not entailing any approximation, all while retaining the linesearch-free nature of adaptive schemes. Furthermore, we prove full sequence convergence without prior knowledge of local H\"older constants nor of the order of H\"older continuity. Numerical experiments make comparisons with baseline methods on diverse tasks from machine learning covering both the locally and the globally H\"older setting.

Related articles: Most relevant | Search more
arXiv:2301.04431 [math.OC] (Published 2023-01-11)
Adaptive proximal algorithms for convex optimization under local Lipschitz continuity of the gradient
arXiv:2311.16850 [math.OC] (Published 2023-11-28)
General Derivative-Free Optimization Methods under Global and Local Lipschitz Continuity of Gradients
arXiv:2408.07182 [math.OC] (Published 2024-08-13)
Proximal random reshuffling under local Lipschitz continuity