arXiv Analytics

Sign in

arXiv:2203.10789 [cs.LG]AbstractReferencesReviewsResources

Domain Generalization by Mutual-Information Regularization with Pre-trained Models

Junbum Cha, Kyungjae Lee, Sungrae Park, Sanghyuk Chun

Published 2022-03-21Version 1

Domain generalization (DG) aims to learn a generalized model to an unseen target domain using only limited source domains. Previous attempts to DG fail to learn domain-invariant representations only from the source domains due to the significant domain shifts between training and test domains. Instead, we re-formulate the DG objective using mutual information with the oracle model, a model generalized to any possible domain. We derive a tractable variational lower bound via approximating the oracle model by a pre-trained model, called Mutual Information Regularization with Oracle (MIRO). Our extensive experiments show that MIRO significantly improves the out-of-distribution performance. Furthermore, our scaling experiments show that the larger the scale of the pre-trained model, the greater the performance improvement of MIRO. Source code is available at https://github.com/kakaobrain/miro.

Related articles: Most relevant | Search more
arXiv:2112.03676 [cs.LG] (Published 2021-12-07, updated 2023-09-17)
PLACE dropout: A Progressive Layer-wise and Channel-wise Dropout for Domain Generalization
arXiv:2109.04320 [cs.LG] (Published 2021-09-09)
COLUMBUS: Automated Discovery of New Multi-Level Features for Domain Generalization via Knowledge Corruption
arXiv:2410.17020 [cs.LG] (Published 2024-10-22)
LFME: A Simple Framework for Learning from Multiple Experts in Domain Generalization