Subsampled Stochastic Variance-Reduced Gradient Langevin Dynamics
Abstract
Stochastic variance-reduced gradient Langevin dynamics (SVRG-LD) was recently proposed to improve the performance of stochastic gradient Langevin dynamics (SGLD) by reducing the variance of the stochastic gradient. In this paper, we propose a variant of SVRG-LD, namely SVRG-LD$^+$, which replaces the full gradient in each epoch with a subsampled one. We provide a nonasymptotic analysis of the convergence of SVRG-LD$^+$ in $2$-Wasserstein distance, and show that SVRG-LD$^+$ enjoys a lower gradient complexity\footnote{Gradient complexity is defined as the required number of stochastic gradient evaluations to reach a target accuracy.} than SVRG-LD, when the sample size is large or the target accuracy requirement is moderate. Our analysis directly implies a sharper convergence rate for SVRG-LD, which improves the existing convergence rate by a factor of $\kappa^{1/6}n^{1/6}$, where $\kappa$ is the condition number of the log-density function and $n$ is the sample size. Experiments on both synthetic and real-world datasets validate our theoretical results.
Cite
Text
Zou et al. "Subsampled Stochastic Variance-Reduced Gradient Langevin Dynamics." Conference on Uncertainty in Artificial Intelligence, 2018.Markdown
[Zou et al. "Subsampled Stochastic Variance-Reduced Gradient Langevin Dynamics." Conference on Uncertainty in Artificial Intelligence, 2018.](https://mlanthology.org/uai/2018/zou2018uai-subsampled/)BibTeX
@inproceedings{zou2018uai-subsampled,
title = {{Subsampled Stochastic Variance-Reduced Gradient Langevin Dynamics}},
author = {Zou, Difan and Xu, Pan and Gu, Quanquan},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {2018},
pages = {508-518},
url = {https://mlanthology.org/uai/2018/zou2018uai-subsampled/}
}