Sinkhorn Barycenter via Functional Gradient Descent
Abstract
In this paper, we consider the problem of computing the barycenter of a set of probability distributions under the Sinkhorn divergence. This problem has recently found applications across various domains, including graphics, learning, and vision, as it provides a meaningful mechanism to aggregate knowledge. Unlike previous approaches which directly operate in the space of probability measures, we recast the Sinkhorn barycenter problem as an instance of unconstrained functional optimization and develop a novel functional gradient descent method named \texttt{Sinkhorn Descent} (\texttt{SD}). We prove that \texttt{SD} converges to a stationary point at a sublinear rate, and under reasonable assumptions, we further show that it asymptotically finds a global minimizer of the Sinkhorn barycenter problem. Moreover, by providing a mean-field analysis, we show that \texttt{SD} preserves the weak convergence of empirical measures. Importantly, the computational complexity of \texttt{SD} scales linearly in the dimension $d$ and we demonstrate its scalability by solving a $100$-dimensional Sinkhorn barycenter problem.
Cite
Text
Shen et al. "Sinkhorn Barycenter via Functional Gradient Descent." Neural Information Processing Systems, 2020.Markdown
[Shen et al. "Sinkhorn Barycenter via Functional Gradient Descent." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/shen2020neurips-sinkhorn/)BibTeX
@inproceedings{shen2020neurips-sinkhorn,
title = {{Sinkhorn Barycenter via Functional Gradient Descent}},
author = {Shen, Zebang and Wang, Zhenfu and Ribeiro, Alejandro and Hassani, Hamed},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/shen2020neurips-sinkhorn/}
}