Stochastic Variance-Reduced Cubic Regularized Newton Methods

Abstract

We propose a stochastic variance-reduced cubic regularized Newton method (SVRC) for non-convex optimization. At the core of our algorithm is a novel semi-stochastic gradient along with a semi-stochastic Hessian, which are specifically designed for cubic regularization method. We show that our algorithm is guaranteed to converge to an $(\epsilon,\sqrt{\epsilon})$-approximate local minimum within $\tilde{O}(n^{4/5}/\epsilon^{3/2})$ second-order oracle calls, which outperforms the state-of-the-art cubic regularization algorithms including subsampled cubic regularization. Our work also sheds light on the application of variance reduction technique to high-order non-convex optimization methods. Thorough experiments on various non-convex optimization problems support our theory.

Cite

Text

Zhou et al. "Stochastic Variance-Reduced Cubic Regularized Newton Methods." International Conference on Machine Learning, 2018.

Markdown

[Zhou et al. "Stochastic Variance-Reduced Cubic Regularized Newton Methods." International Conference on Machine Learning, 2018.](https://mlanthology.org/icml/2018/zhou2018icml-stochastic/)

BibTeX

@inproceedings{zhou2018icml-stochastic,
  title     = {{Stochastic Variance-Reduced Cubic Regularized Newton Methods}},
  author    = {Zhou, Dongruo and Xu, Pan and Gu, Quanquan},
  booktitle = {International Conference on Machine Learning},
  year      = {2018},
  pages     = {5990-5999},
  volume    = {80},
  url       = {https://mlanthology.org/icml/2018/zhou2018icml-stochastic/}
}