Wasserstein Distributional Learning via Majorization-Minimization

Abstract

Learning function-on-scalar predictive models for conditional densities and identifying factors that influence the entire probability distribution are vital tasks in many data-driven applications. We present an efficient Majorization-Minimization optimization algorithm, Wasserstein Distributional Learning (WDL), that trains Semi-parametric Conditional Gaussian Mixture Models (SCGMM) for conditional density functions and uses the Wasserstein distance $W_2$ as a proper metric for the space of density outcomes. We further provide theoretical convergence guarantees and illustrate the algorithm using boosted machines. Experiments on the synthetic data and real-world applications demonstrate the effectiveness of the proposed WDL algorithm.

Cite

Text

Tang et al. "Wasserstein Distributional Learning via Majorization-Minimization." Artificial Intelligence and Statistics, 2023.

Markdown

[Tang et al. "Wasserstein Distributional Learning via Majorization-Minimization." Artificial Intelligence and Statistics, 2023.](https://mlanthology.org/aistats/2023/tang2023aistats-wasserstein/)

BibTeX

@inproceedings{tang2023aistats-wasserstein,
  title     = {{Wasserstein Distributional Learning via Majorization-Minimization}},
  author    = {Tang, Chengliang and Lenssen, Nathan and Wei, Ying and Zheng, Tian},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2023},
  pages     = {10703-10731},
  volume    = {206},
  url       = {https://mlanthology.org/aistats/2023/tang2023aistats-wasserstein/}
}