arXiv Analytics

Sign in

arXiv:2304.10552 [cs.LG]AbstractReferencesReviewsResources

Interpolation property of shallow neural networks

Vlad-Raul Constantinescu, Ionel Popescu

Published 2023-04-20Version 1

We study the geometry of global minima of the loss landscape of overparametrized neural networks. In most optimization problems, the loss function is convex, in which case we only have a global minima, or nonconvex, with a discrete number of global minima. In this paper, we prove that in the overparametrized regime, a shallow neural network can interpolate any data set, i.e. the loss function has a global minimum value equal to zero as long as the activation function is not a polynomial of small degree. Additionally, if such a global minimum exists, then the locus of global minima has infinitely many points. Furthermore, we give a characterization of the Hessian of the loss function evaluated at the global minima, and in the last section, we provide a practical probabilistic method of finding the interpolation point.

Related articles: Most relevant | Search more
arXiv:1905.13200 [cs.LG] (Published 2019-05-30)
Exploiting Uncertainty of Loss Landscape for Stochastic Optimization
arXiv:1903.02893 [cs.LG] (Published 2019-03-07)
Only sparsity based loss function for learning representations
arXiv:1901.09178 [cs.LG] (Published 2019-01-26)
A general model for plane-based clustering with loss function