arXiv Analytics

Sign in

arXiv:1511.07948 [cs.LG]AbstractReferencesReviewsResources

Learning Halfspaces and Neural Networks with Random Initialization

Yuchen Zhang, Jason D. Lee, Martin J. Wainwright, Michael I. Jordan

Published 2015-11-25Version 1

We study non-convex empirical risk minimization for learning halfspaces and neural networks. For loss functions that are $L$-Lipschitz continuous, we present algorithms to learn halfspaces and multi-layer neural networks that achieve arbitrarily small excess risk $\epsilon>0$. The time complexity is polynomial in the input dimension $d$ and the sample size $n$, but exponential in the quantity $(L/\epsilon^2)\log(L/\epsilon)$. These algorithms run multiple rounds of random initialization followed by arbitrary optimization steps. We further show that if the data is separable by some neural network with constant margin $\gamma>0$, then there is a polynomial-time algorithm for learning a neural network that separates the training data with margin $\Omega(\gamma)$. As a consequence, the algorithm achieves arbitrary generalization error $\epsilon>0$ with ${\rm poly}(d,1/\epsilon)$ sample and time complexity. We establish the same learnability result when the labels are randomly flipped with probability $\eta<1/2$.

Related articles: Most relevant | Search more
arXiv:1801.00711 [cs.LG] (Published 2017-12-25)
Network-Scale Traffic Modeling and Forecasting with Graphical Lasso and Neural Networks
arXiv:1905.06549 [cs.LG] (Published 2019-05-16)
TapNet: Neural Network Augmented with Task-Adaptive Projection for Few-Shot Learning
arXiv:2012.10985 [cs.LG] (Published 2020-12-20)
Learning Halfspaces With Membership Queries