arXiv Analytics

Sign in

arXiv:2211.08253 [cs.LG]AbstractReferencesReviewsResources

HMOE: Hypernetwork-based Mixture of Experts for Domain Generalization

Jingang Qu, Thibault Faney, Ze Wang, Patrick Gallinari, Soleiman Yousef, Jean-Charles de Hemptinne

Published 2022-11-15Version 1

Due to the domain shift, machine learning systems typically fail to generalize well to domains different from those of training data, which is the problem that domain generalization (DG) aims to address. However, most mainstream DG algorithms lack interpretability and require domain labels, which are not available in many real-world scenarios. In this work, we propose a novel DG method, HMOE: Hypernetwork-based Mixture of Experts (MoE), that does not require domain labels and is more interpretable. We use hypernetworks to generate the weights of experts, allowing experts to share some useful meta-knowledge. MoE has proven adept at detecting and identifying heterogeneous patterns in data. For DG, heterogeneity exactly arises from the domain shift. We compare HMOE with other DG algorithms under a fair and unified benchmark-DomainBed. Extensive experiments show that HMOE can perform latent domain discovery from data of mixed domains and divide it into distinct clusters that are surprisingly more consistent with human intuition than original domain labels. Compared to other DG methods, HMOE shows competitive performance and achieves SOTA results in some cases without using domain labels.

Related articles: Most relevant | Search more
arXiv:2410.06020 [cs.LG] (Published 2024-10-08)
QT-DoG: Quantization-aware Training for Domain Generalization
arXiv:2107.10931 [cs.LG] (Published 2021-07-22)
Domain Generalization under Conditional and Label Shifts via Variational Bayesian Inference
Xiaofeng Liu et al.
arXiv:2404.02785 [cs.LG] (Published 2024-04-03)
Domain Generalization through Meta-Learning: A Survey