High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm
Abstract
We propose a generic stochastic expectation-maximization (EM) algorithm for the estimation of high-dimensional latent variable models. At the core of our algorithm is a novel semi-stochastic variance-reduced gradient designed for the $Q$-function in the EM algorithm. Under a mild condition on the initialization, our algorithm is guaranteed to attain a linear convergence rate to the unknown parameter of the latent variable model, and achieve an optimal statistical rate up to a logarithmic factor for parameter estimation. Compared with existing high-dimensional EM algorithms, our algorithm enjoys a better computational complexity and is therefore more efficient. We apply our generic algorithm to two illustrative latent variable models: Gaussian mixture model and mixture of linear regression, and demonstrate the advantages of our algorithm by both theoretical analysis and numerical experiments. We believe that the proposed semi-stochastic gradient is of independent interest for general nonconvex optimization problems with bivariate structures.
Cite
Text
Zhu et al. "High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm." International Conference on Machine Learning, 2017.Markdown
[Zhu et al. "High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm." International Conference on Machine Learning, 2017.](https://mlanthology.org/icml/2017/zhu2017icml-highdimensional/)BibTeX
@inproceedings{zhu2017icml-highdimensional,
title = {{High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm}},
author = {Zhu, Rongda and Wang, Lingxiao and Zhai, Chengxiang and Gu, Quanquan},
booktitle = {International Conference on Machine Learning},
year = {2017},
pages = {4180-4188},
volume = {70},
url = {https://mlanthology.org/icml/2017/zhu2017icml-highdimensional/}
}