arXiv Analytics

Sign in

arXiv:1710.07138 [stat.ML]AbstractReferencesReviewsResources

Binary Classification from Positive-Confidence Data

Takashi Ishida, Gang Niu, Masashi Sugiyama

Published 2017-10-19Version 1

Reducing labeling costs in supervised learning is a critical issue in many practical machine learning applications. In this paper, we consider positive-confidence (Pconf) classification, the problem of training a binary classifier only from positive data equipped with confidence. Pconf classification can be regarded as a discriminative extension of one-class classification (which is aimed at "describing" the positive class), with ability to tune hyper-parameters for "classifying" positive and negative samples. Pconf classification is also related to positive-unlabeled (PU) classification (which uses hard-labeled positive data and unlabeled data), allowing us to avoid estimating the class priors, which is a critical bottleneck in typical PU classification methods. For the Pconf classification problem, we provide a simple empirical risk minimization framework and give a formulation for linear-in-parameter models that can be implemented easily and computationally efficiently. We also theoretically establish the consistency and generalization error bounds for Pconf classification, and demonstrate the practical usefulness of the proposed method through experiments.

Related articles: Most relevant | Search more
arXiv:2201.11059 [stat.ML] (Published 2021-12-23)
Generalization Error Bounds on Deep Learning with Markov Datasets
arXiv:2308.01835 [stat.ML] (Published 2023-08-03)
Distribution-Free Inference for the Regression Function of Binary Classification
arXiv:1806.00640 [stat.ML] (Published 2018-06-02)
Binary Classification with Karmic, Threshold-Quasi-Concave Metrics