arXiv Analytics

Sign in

arXiv:2302.03115 [cs.LG]AbstractReferencesReviewsResources

Easy Learning from Label Proportions

Robert Istvan Busa-Fekete, Heejin Choi, Travis Dick, Claudio Gentile, Andres Munoz medina

Published 2023-02-06Version 1

We consider the problem of Learning from Label Proportions (LLP), a weakly supervised classification setup where instances are grouped into "bags", and only the frequency of class labels at each bag is available. Albeit, the objective of the learner is to achieve low task loss at an individual instance level. Here we propose Easyllp: a flexible and simple-to-implement debiasing approach based on aggregate labels, which operates on arbitrary loss functions. Our technique allows us to accurately estimate the expected loss of an arbitrary model at an individual level. We showcase the flexibility of our approach by applying it to popular learning frameworks, like Empirical Risk Minimization (ERM) and Stochastic Gradient Descent (SGD) with provable guarantees on instance level performance. More concretely, we exhibit a variance reduction technique that makes the quality of LLP learning deteriorate only by a factor of k (k being bag size) in both ERM and SGD setups, as compared to full supervision. Finally, we validate our theoretical results on multiple datasets demonstrating our algorithm performs as well or better than previous LLP approaches in spite of its simplicity.

Related articles: Most relevant | Search more
arXiv:1905.12909 [cs.LG] (Published 2019-05-30)
Deep multi-class learning from label proportions
arXiv:1909.02180 [cs.LG] (Published 2019-09-05)
Learning from Label Proportions with Generative Adversarial Networks
arXiv:2004.03515 [cs.LG] (Published 2020-04-07)
On the Complexity of Learning from Label Proportions