arXiv Analytics

Sign in

arXiv:2007.06168 [cs.LG]AbstractReferencesReviewsResources

Model Fusion with Kullback--Leibler Divergence

Sebastian Claici, Mikhail Yurochkin, Soumya Ghosh, Justin Solomon

Published 2020-07-13Version 1

We propose a method to fuse posterior distributions learned from heterogeneous datasets. Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors and proceeds using a simple assign-and-average approach. The components of the dataset posteriors are assigned to the proposed global model components by solving a regularized variant of the assignment problem. The global components are then updated based on these assignments by their mean under a KL divergence. For exponential family variational distributions, our formulation leads to an efficient non-parametric algorithm for computing the fused model. Our algorithm is easy to describe and implement, efficient, and competitive with state-of-the-art on motion capture analysis, topic modeling, and federated learning of Bayesian neural networks.

Related articles: Most relevant | Search more
arXiv:2502.00264 [cs.LG] (Published 2025-02-01)
Beyond the Permutation Symmetry of Transformers: The Role of Rotation for Model Fusion
arXiv:1606.05850 [cs.LG] (Published 2016-06-19)
Guaranteed bounds on the Kullback-Leibler divergence of univariate mixtures using piecewise log-sum-exp inequalities
arXiv:1004.5229 [cs.LG] (Published 2010-04-29, updated 2010-10-13)
Optimism in Reinforcement Learning and Kullback-Leibler Divergence