arXiv Analytics

Sign in

arXiv:1905.13736 [stat.ML]AbstractReferencesReviewsResources

Unlabeled Data Improves Adversarial Robustness

Yair Carmon, Aditi Raghunathan, Ludwig Schmidt, Percy Liang, John C. Duchi

Published 2019-05-31Version 1

We demonstrate, theoretically and empirically, that adversarial robustness can significantly benefit from semisupervised learning. Theoretically, we revisit the simple Gaussian model of Schmidt et al. that shows a sample complexity gap between standard and robust classification. We prove that this gap does not pertain to labels: a simple semisupervised learning procedure (self-training) achieves robust accuracy using the same number of labels required for standard accuracy. Empirically, we augment CIFAR-10 with 500K unlabeled images sourced from 80 Million Tiny Images and use robust self-training to outperform state-of-the-art robust accuracies by over 5 points in (i) $\ell_\infty$ robustness against several strong attacks via adversarial training and (ii) certified $\ell_2$ and $\ell_\infty$ robustness via randomized smoothing. On SVHN, adding the dataset's own extra training set with the labels removed provides gains of 4 to 10 points, within 1 point of the gain from using the extra labels as well.

Related articles: Most relevant | Search more
arXiv:1805.12152 [stat.ML] (Published 2018-05-30)
There Is No Free Lunch In Adversarial Robustness (But There Are Unexpected Benefits)
arXiv:2502.01027 [stat.ML] (Published 2025-02-03)
Adversarial Robustness in Two-Stage Learning-to-Defer: Algorithms and Guarantees
arXiv:1909.08079 [stat.ML] (Published 2019-09-17)
Relaxed Softmax for learning from Positive and Unlabeled data