arXiv Analytics

Sign in

arXiv:1810.09038 [stat.ML]AbstractReferencesReviewsResources

Depth with Nonlinearity Creates No Bad Local Minima in ResNets

Kenji Kawaguchi, Yoshua Bengio

Published 2018-10-21Version 1

In this paper, we prove that depth with nonlinearity creates no bad local minima in a type of arbitrarily deep ResNets studied in previous work, in the sense that the values of all local minima are no worse than the global minima values of corresponding shallow linear predictors with arbitrary fixed features, and are guaranteed to further improve via residual representations. As a result, this paper provides an affirmative answer to an open question stated in a paper in the conference on Neural Information Processing Systems (NIPS) 2018. We note that even though our paper advances the theoretical foundation of deep learning and non-convex optimization, there is still a gap between theory and many practical deep learning applications.

Related articles: Most relevant | Search more
arXiv:1901.03909 [stat.ML] (Published 2019-01-12)
Eliminating all bad Local Minima from Loss Landscapes without even adding an Extra Unit
arXiv:1611.06310 [stat.ML] (Published 2016-11-19)
Local minima in training of deep networks
arXiv:1805.08671 [stat.ML] (Published 2018-05-22)
Adding One Neuron Can Eliminate All Bad Local Minima