Model Fusion with Kullback-Leibler Divergence

Abstract

We propose a method to fuse posterior distributions learned from heterogeneous datasets. Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors and proceeds using a simple assign-and-average approach. The components of the dataset posteriors are assigned to the proposed global model components by solving a regularized variant of the assignment problem. The global components are then updated based on these assignments by their mean under a KL divergence. For exponential family variational distributions, our formulation leads to an efficient non-parametric algorithm for computing the fused model. Our algorithm is easy to describe and implement, efficient, and competitive with state-of-the-art on motion capture analysis, topic modeling, and federated learning of Bayesian neural networks.

Cite

Text

Claici et al. "Model Fusion with Kullback-Leibler Divergence." International Conference on Machine Learning, 2020.

Markdown

[Claici et al. "Model Fusion with Kullback-Leibler Divergence." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/claici2020icml-model/)

BibTeX

@inproceedings{claici2020icml-model,
  title     = {{Model Fusion with Kullback-Leibler Divergence}},
  author    = {Claici, Sebastian and Yurochkin, Mikhail and Ghosh, Soumya and Solomon, Justin},
  booktitle = {International Conference on Machine Learning},
  year      = {2020},
  pages     = {2038-2047},
  volume    = {119},
  url       = {https://mlanthology.org/icml/2020/claici2020icml-model/}
}