arXiv Analytics

Sign in

arXiv:2412.14031 [math.OC]AbstractReferencesReviewsResources

Gauss-Newton Dynamics for Neural Networks: A Riemannian Optimization Perspective

Semih Cayci

Published 2024-12-18Version 1

We analyze the convergence of Gauss-Newton dynamics for training neural networks with smooth activation functions. In the underparameterized regime, the Gauss-Newton gradient flow induces a Riemannian gradient flow on a low-dimensional, smooth, embedded submanifold of the Euclidean output space. Using tools from Riemannian optimization, we prove \emph{last-iterate} convergence of the Riemannian gradient flow to the optimal in-class predictor at an \emph{exponential rate} that is independent of the conditioning of the Gram matrix, \emph{without} requiring explicit regularization. We further characterize the critical impacts of the neural network scaling factor and the initialization on the convergence behavior. In the overparameterized regime, we show that the Levenberg-Marquardt dynamics with an appropriately chosen damping factor yields robustness to ill-conditioned kernels, analogous to the underparameterized regime. These findings demonstrate the potential of Gauss-Newton methods for efficiently optimizing neural networks, particularly in ill-conditioned problems where kernel and Gram matrices have small singular values.

Related articles: Most relevant | Search more
arXiv:1910.01895 [math.OC] (Published 2019-10-04)
Approximate policy iteration using neural networks for storage problems
arXiv:2303.04436 [math.OC] (Published 2023-03-08, updated 2023-09-07)
A comparison of rational and neural network based approximations
arXiv:1802.08539 [math.OC] (Published 2018-02-23)
Computation of optimal transport and related hedging problems via penalization and neural networks