An Accelerated Variance Reducing Stochastic Method with Douglas-Rachford Splitting

Abstract

We consider the problem of minimizing the regularized empirical risk function which is represented as the average of a large number of convex loss functions plus a possibly non-smooth convex regularization term. In this paper, we propose a fast variance reducing (VR) stochastic method called Prox2-SAGA. Different from traditional VR stochastic methods, Prox2-SAGA replaces the stochastic gradient of the loss function with the corresponding gradient mapping. In addition, Prox2-SAGA also computes the gradient mapping of the regularization term. These two gradient mappings constitute a Douglas-Rachford splitting step. For strongly convex and smooth loss functions, we prove that Prox2-SAGA can achieve a linear convergence rate comparable to other accelerated VR stochastic methods. In addition, Prox2-SAGA is more practical as it involves only the stepsize to tune. When each loss function is smooth but non-strongly convex, we prove a convergence rate of ${\mathcal {O}}(1/k)$ O ( 1 / k ) for the proposed Prox2-SAGA method, where k is the number of iterations. Moreover, experiments show that Prox2-SAGA is valid for non-smooth loss functions, and for strongly convex and smooth loss functions, Prox2-SAGA is prominently faster when loss functions are ill-conditioned.

Cite

Text

Liu et al. "An Accelerated Variance Reducing Stochastic Method with Douglas-Rachford Splitting." Machine Learning, 2019. doi:10.1007/S10994-019-05785-3

Markdown

[Liu et al. "An Accelerated Variance Reducing Stochastic Method with Douglas-Rachford Splitting." Machine Learning, 2019.](https://mlanthology.org/mlj/2019/liu2019mlj-accelerated/) doi:10.1007/S10994-019-05785-3

BibTeX

@article{liu2019mlj-accelerated,
  title     = {{An Accelerated Variance Reducing Stochastic Method with Douglas-Rachford Splitting}},
  author    = {Liu, Jingchang and Xu, Linli and Shen, Shuheng and Ling, Qing},
  journal   = {Machine Learning},
  year      = {2019},
  pages     = {859-878},
  doi       = {10.1007/S10994-019-05785-3},
  volume    = {108},
  url       = {https://mlanthology.org/mlj/2019/liu2019mlj-accelerated/}
}