arXiv Analytics

Sign in

arXiv:2006.06467 [cs.LG]AbstractReferencesReviewsResources

Learning Halfspaces with Tsybakov Noise

Ilias Diakonikolas, Vasilis Kontonis, Christos Tzamos, Nikos Zarifis

Published 2020-06-11Version 1

We study the efficient PAC learnability of halfspaces in the presence of Tsybakov noise. In the Tsybakov noise model, each label is independently flipped with some probability which is controlled by an adversary. This noise model significantly generalizes the Massart noise model, by allowing the flipping probabilities to be arbitrarily close to $1/2$ for a fraction of the samples. Our main result is the first non-trivial PAC learning algorithm for this problem under a broad family of structured distributions -- satisfying certain concentration and (anti-)anti-concentration properties -- including log-concave distributions. Specifically, we given an algorithm that achieves misclassification error $\epsilon$ with respect to the true halfspace, with quasi-polynomial runtime dependence in $1/\epsilin$. The only previous upper bound for this problem -- even for the special case of log-concave distributions -- was doubly exponential in $1/\epsilon$ (and follows via the naive reduction to agnostic learning). Our approach relies on a novel computationally efficient procedure to certify whether a candidate solution is near-optimal, based on semi-definite programming. We use this certificate procedure as a black-box and turn it into an efficient learning algorithm by searching over the space of halfspaces via online convex optimization.

Related articles: Most relevant | Search more
arXiv:2108.08767 [cs.LG] (Published 2021-08-19)
Threshold Phenomena in Learning Halfspaces with Massart Noise
arXiv:2012.10985 [cs.LG] (Published 2020-12-20)
Learning Halfspaces With Membership Queries
arXiv:2102.09305 [cs.LG] (Published 2021-02-18)
Boosting for Online Convex Optimization