Stochastic Optimization with Importance Sampling for Regularized Loss Minimization
Abstract
Uniform sampling of training data has been commonly used in traditional stochastic optimization algorithms such as Proximal Stochastic Mirror Descent (prox-SMD) and Proximal Stochastic Dual Coordinate Ascent (prox-SDCA). Although uniform sampling can guarantee that the sampled stochastic quantity is an unbiased estimate of the corresponding true quantity, the resulting estimator may have a rather high variance, which negatively affects the convergence of the underlying optimization procedure. In this paper we study stochastic optimization, including prox-SMD and prox-SDCA, with importance sampling, which improves the convergence rate by reducing the stochastic variance. We theoretically analyze the algorithms and empirically validate their effectiveness.
Cite
Text
Zhao and Zhang. "Stochastic Optimization with Importance Sampling for Regularized Loss Minimization." International Conference on Machine Learning, 2015.Markdown
[Zhao and Zhang. "Stochastic Optimization with Importance Sampling for Regularized Loss Minimization." International Conference on Machine Learning, 2015.](https://mlanthology.org/icml/2015/zhao2015icml-stochastic/)BibTeX
@inproceedings{zhao2015icml-stochastic,
title = {{Stochastic Optimization with Importance Sampling for Regularized Loss Minimization}},
author = {Zhao, Peilin and Zhang, Tong},
booktitle = {International Conference on Machine Learning},
year = {2015},
pages = {1-9},
volume = {37},
url = {https://mlanthology.org/icml/2015/zhao2015icml-stochastic/}
}