arXiv Analytics

Sign in

arXiv:1901.09960 [cs.LG]AbstractReferencesReviewsResources

Using Pre-Training Can Improve Model Robustness and Uncertainty

Dan Hendrycks, Kimin Lee, Mantas Mazeika

Published 2019-01-28Version 1

Tuning a pre-trained network is commonly thought to improve data efficiency. However, Kaiming He et al. have called into question the utility of pre-training by showing that training from scratch can often yield similar performance, should the model train long enough. We show that although pre-training may not improve performance on traditional classification metrics, it does provide large benefits to model robustness and uncertainty. Through extensive experiments on label corruption, class imbalance, adversarial examples, out-of-distribution detection, and confidence calibration, we demonstrate large gains from pre-training and complementary effects with task-specific methods. We show approximately a 30% relative improvement in label noise robustness and a 10% absolute improvement in adversarial robustness on CIFAR-10 and CIFAR-100. In some cases, using pre-training without task-specific methods surpasses the state-of-the-art, highlighting the importance of using pre-training when evaluating future methods on robustness and uncertainty tasks.

Related articles: Most relevant | Search more
arXiv:1906.12340 [cs.LG] (Published 2019-06-28)
Using Self-Supervised Learning Can Improve Model Robustness and Uncertainty
arXiv:2002.05380 [cs.LG] (Published 2020-02-13)
CEB Improves Model Robustness
arXiv:2308.08934 [cs.LG] (Published 2023-08-17)
On Data Imbalance in Molecular Property Prediction with Pre-training