arXiv Analytics

Sign in

arXiv:2004.03515 [cs.LG]AbstractReferencesReviewsResources

On the Complexity of Learning from Label Proportions

Benjamin Fish, Lev Reyzin

Published 2020-04-07Version 1

In the problem of learning with label proportions, which we call LLP learning, the training data is unlabeled, and only the proportions of examples receiving each label are given. The goal is to learn a hypothesis that predicts the proportions of labels on the distribution underlying the sample. This model of learning is applicable to a wide variety of settings, including predicting the number of votes for candidates in political elections from polls. In this paper, we formally define this class and resolve foundational questions regarding the computational complexity of LLP and characterize its relationship to PAC learning. Among our results, we show, perhaps surprisingly, that for finite VC classes what can be efficiently LLP learned is a strict subset of what can be leaned efficiently in PAC, under standard complexity assumptions. We also show that there exist classes of functions whose learnability in LLP is independent of ZFC, the standard set theoretic axioms. This implies that LLP learning cannot be easily characterized (like PAC by VC dimension).

Comments: this is an extended and corrected version of an IJCAI 2017 paper, 13 pages
Categories: cs.LG, stat.ML
Related articles: Most relevant | Search more
arXiv:2302.03115 [cs.LG] (Published 2023-02-06)
Easy Learning from Label Proportions
arXiv:1905.12909 [cs.LG] (Published 2019-05-30)
Deep multi-class learning from label proportions
arXiv:1909.02180 [cs.LG] (Published 2019-09-05)
Learning from Label Proportions with Generative Adversarial Networks