arXiv Analytics

Sign in

arXiv:2012.04550 [cs.LG]AbstractReferencesReviewsResources

In-N-Out: Pre-Training and Self-Training using Auxiliary Information for Out-of-Distribution Robustness

Sang Michael Xie, Ananya Kumar, Robbie Jones, Fereshte Khani, Tengyu Ma, Percy Liang

Published 2020-12-08Version 1

Consider a prediction setting where a few inputs (e.g., satellite images) are expensively annotated with the prediction targets (e.g., crop types), and many inputs are cheaply annotated with auxiliary information (e.g., climate information). How should we best leverage this auxiliary information for the prediction task? Empirically across three image and time-series datasets, and theoretically in a multi-task linear regression setting, we show that (i) using auxiliary information as input features improves in-distribution error but can hurt out-of-distribution (OOD) error; while (ii) using auxiliary information as outputs of auxiliary tasks to pre-train a model improves OOD error. To get the best of both worlds, we introduce In-N-Out, which first trains a model with auxiliary inputs and uses it to pseudolabel all the in-distribution inputs, then pre-trains a model on OOD auxiliary outputs and fine-tunes this model with the pseudolabels (self-training). We show both theoretically and empirically that In-N-Out outperforms auxiliary inputs or outputs alone on both in-distribution and OOD error.

Related articles: Most relevant | Search more
arXiv:2402.19102 [cs.LG] (Published 2024-02-29)
FlatNAS: optimizing Flatness in Neural Architecture Search for Out-of-Distribution Robustness
arXiv:2107.07564 [cs.LG] (Published 2021-07-15)
On the Importance of Regularisation & Auxiliary Information in OOD Detection
arXiv:2206.00395 [cs.LG] (Published 2022-06-01)
Optimization with access to auxiliary information