Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections
Abstract
We consider stochastic strongly convex optimization with a complex inequality constraint. This complex inequality constraint may lead to computationally expensive projections in algorithmic iterations of the stochastic gradient descent~(SGD) methods. To reduce the computation costs pertaining to the projections, we propose an Epoch-Projection Stochastic Gradient Descent~(Epro-SGD) method. The proposed Epro-SGD method consists of a sequence of epochs; it applies SGD to an augmented objective function at each iteration within the epoch, and then performs a projection at the end of each epoch. Given a strongly convex optimization and for a total number of $T$ iterations, Epro-SGD requires only $\log(T)$ projections, and meanwhile attains an optimal convergence rate of $O(1/T)$, both in expectation and with a high probability. To exploit the structure of the optimization problem, we propose a proximal variant of Epro-SGD, namely Epro-ORDA, based on the optimal regularized dual averaging method. We apply the proposed methods on real-world applications; the empirical results demonstrate the effectiveness of our methods.
Cite
Text
Chen et al. "Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections." Conference on Uncertainty in Artificial Intelligence, 2016.Markdown
[Chen et al. "Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections." Conference on Uncertainty in Artificial Intelligence, 2016.](https://mlanthology.org/uai/2016/chen2016uai-optimal/)BibTeX
@inproceedings{chen2016uai-optimal,
title = {{Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections}},
author = {Chen, Jianhui and Yang, Tianbao and Lin, Qihang and Zhang, Lijun and Chang, Yi},
booktitle = {Conference on Uncertainty in Artificial Intelligence},
year = {2016},
url = {https://mlanthology.org/uai/2016/chen2016uai-optimal/}
}