Finite-Sum Composition Optimization via Variance Reduced Gradient Descent

Abstract

The stochastic composition optimization proposed recently by Wang et al. [2014] minimizes the objective with the compositional expectation form: $\min_x~(\mathbb{E}_iF_i \circ \mathbb{E}_j G_j)(x).$ It summarizes many important applications in machine learning, statistics, and finance. In this paper, we consider the finite-sum scenario for composition optimization: \[\min_x f (x) := \frac{1}n \sum_i = 1^n F_i \left(\frac{1}m \sum_j = 1^m G_j (x) \right). \] We propose two algorithms to solve this problem by combining the stochastic compositional gradient descent (SCGD) and the stochastic variance reduced gradient (SVRG) technique. A constant linear convergence rate is proved for strongly convex optimization, which substantially improves the sublinear rate $O(K^{-0.8})$ of the best known algorithm.

Cite

Text

Lian et al. "Finite-Sum Composition Optimization via Variance Reduced Gradient Descent." International Conference on Artificial Intelligence and Statistics, 2017.

Markdown

[Lian et al. "Finite-Sum Composition Optimization via Variance Reduced Gradient Descent." International Conference on Artificial Intelligence and Statistics, 2017.](https://mlanthology.org/aistats/2017/lian2017aistats-finite/)

BibTeX

@inproceedings{lian2017aistats-finite,
  title     = {{Finite-Sum Composition Optimization via Variance Reduced Gradient Descent}},
  author    = {Lian, Xiangru and Wang, Mengdi and Liu, Ji},
  booktitle = {International Conference on Artificial Intelligence and Statistics},
  year      = {2017},
  pages     = {1159-1167},
  url       = {https://mlanthology.org/aistats/2017/lian2017aistats-finite/}
}