Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections

Abstract

We consider stochastic strongly convex optimization with a complex inequality constraint. This complex inequality constraint may lead to computationally expensive projections in algorithmic iterations of the stochastic gradient descent~(SGD) methods. To reduce the computation costs pertaining to the projections, we propose an Epoch-Projection Stochastic Gradient Descent~(Epro-SGD) method. The proposed Epro-SGD method consists of a sequence of epochs; it applies SGD to an augmented objective function at each iteration within the epoch, and then performs a projection at the end of each epoch. Given a strongly convex optimization and for a total number of $T$ iterations, Epro-SGD requires only $\log(T)$ projections, and meanwhile attains an optimal convergence rate of $O(1/T)$, both in expectation and with a high probability. To exploit the structure of the optimization problem, we propose a proximal variant of Epro-SGD, namely Epro-ORDA, based on the optimal regularized dual averaging method. We apply the proposed methods on real-world applications; the empirical results demonstrate the effectiveness of our methods.

Cite

Text

Chen et al. "Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections." Conference on Uncertainty in Artificial Intelligence, 2016.

Markdown

[Chen et al. "Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections." Conference on Uncertainty in Artificial Intelligence, 2016.](https://mlanthology.org/uai/2016/chen2016uai-optimal/)

BibTeX

@inproceedings{chen2016uai-optimal,
  title     = {{Optimal Stochastic Strongly Convex Optimization with a Logarithmic Number of Projections}},
  author    = {Chen, Jianhui and Yang, Tianbao and Lin, Qihang and Zhang, Lijun and Chang, Yi},
  booktitle = {Conference on Uncertainty in Artificial Intelligence},
  year      = {2016},
  url       = {https://mlanthology.org/uai/2016/chen2016uai-optimal/}
}