arXiv Analytics

Sign in

arXiv:2405.16027 [cs.LG]AbstractReferencesReviewsResources

Feature Protection For Out-of-distribution Generalization

Lu Tan, Huei Zhou, Yinxiang Huang, Zeming Zheng, Yujiu Yang

Published 2024-05-25Version 1

With the availability of large pre-trained models, a modern workflow for building real-world machine learning solutions is to fine-tune such models on a downstream task with a relatively small domain-specific dataset. In such applications, one major challenge is that the small fine-tuning dataset does not have sufficient coverage of the distribution encountered when the model is deployed. It is thus important to design fine-tuning methods that are robust to out-of-distribution (OOD) data that are under-represented by the training data. This paper compares common fine-tuning methods to investigate their OOD performance and demonstrates that standard methods will result in a significant change to the pre-trained model so that the fine-tuned features overfit the fine-tuning dataset. However, this causes deteriorated OOD performance. To overcome this issue, we show that protecting pre-trained features leads to a fine-tuned model more robust to OOD generalization. We validate the feature protection methods with extensive experiments of fine-tuning CLIP on ImageNet and DomainNet.

Comments: arXiv admin note: substantial text overlap with arXiv:2309.06256
Categories: cs.LG
Related articles: Most relevant | Search more
arXiv:2010.15775 [cs.LG] (Published 2020-10-29)
Understanding the Failure Modes of Out-of-Distribution Generalization
arXiv:2410.11397 [cs.LG] (Published 2024-10-15)
FOOGD: Federated Collaboration for Both Out-of-distribution Generalization and Detection
Xinting Liao et al.
arXiv:2406.00661 [cs.LG] (Published 2024-06-02)
Bridging Multicalibration and Out-of-distribution Generalization Beyond Covariate Shift