arXiv Analytics

Sign in

arXiv:2004.00968 [math.OC]AbstractReferencesReviewsResources

Using gradient directions to get global convergence of Newton-type methods

Daniela di Serafino, Gerardo Toraldo, Marco Viola

Published 2020-04-02Version 1

The renewed interest in Steepest Descent (SD) methods following the work of Barzilai and Borwein [IMA Journal of Numerical Analysis, 8 (1988)] has driven us to consider a globalization strategy based on SD, which is applicable to any line-search method. In particular, we combine Newton-type directions with scaled SD steps to have suitable descent directions. Scaling the SD directions with a suitable step length makes a significant difference with respect to similar globalization approaches, in terms of both theoretical features and computational behavior. We apply our strategy to Newton's method and the BFGS method, with computational results that appear interesting compared with the results of well-established globalization strategies devised ad hoc for those methods.

Related articles: Most relevant | Search more
arXiv:2202.02914 [math.OC] (Published 2022-02-07)
Global convergence and optimality of the heavy ball method for non-convex optimization
arXiv:2304.09537 [math.OC] (Published 2023-04-19)
Global Convergence of Algorithms Based on Unions of Nonexpansive Maps
arXiv:2407.00667 [math.OC] (Published 2024-06-30)
Gradient directions and relative inexactness in optimization and machine learning