arXiv Analytics

Sign in

arXiv:2408.01872 [cs.LG]AbstractReferencesReviewsResources

Safe Semi-Supervised Contrastive Learning Using In-Distribution Data as Positive Examples

Min Gu Kwak, Hyungu Kahng, Seoung Bum Kim

Published 2024-08-03Version 1

Semi-supervised learning methods have shown promising results in solving many practical problems when only a few labels are available. The existing methods assume that the class distributions of labeled and unlabeled data are equal; however, their performances are significantly degraded in class distribution mismatch scenarios where out-of-distribution (OOD) data exist in the unlabeled data. Previous safe semi-supervised learning studies have addressed this problem by making OOD data less likely to affect training based on labeled data. However, even if the studies effectively filter out the unnecessary OOD data, they can lose the basic information that all data share regardless of class. To this end, we propose to apply a self-supervised contrastive learning approach to fully exploit a large amount of unlabeled data. We also propose a contrastive loss function with coefficient schedule to aggregate as an anchor the labeled negative examples of the same class into positive examples. To evaluate the performance of the proposed method, we conduct experiments on image classification datasets - CIFAR-10, CIFAR-100, Tiny ImageNet, and CIFAR-100+Tiny ImageNet - under various mismatch ratios. The results show that self-supervised contrastive learning significantly improves classification accuracy. Moreover, aggregating the in-distribution examples produces better representation and consequently further improves classification accuracy.

Related articles: Most relevant | Search more
arXiv:1911.05275 [cs.LG] (Published 2019-11-13)
Learning from a Teacher using Unlabeled Data
arXiv:2007.01293 [cs.LG] (Published 2020-07-02)
Not All Unlabeled Data are Equal: Learning to Weight Data in Semi-supervised Learning
arXiv:1911.08696 [cs.LG] (Published 2019-11-20)
Where is the Bottleneck of Adversarial Learning with Unlabeled Data?