SPAN: A Stochastic Projected Approximate Newton Method

Abstract

Second-order optimization methods have desirable convergence properties. However, the exact Newton method requires expensive computation for the Hessian and its inverse. In this paper, we propose SPAN, a novel approximate and fast Newton method. SPAN computes the inverse of the Hessian matrix via low-rank approximation and stochastic Hessian-vector products. Our experiments on multiple benchmark datasets demonstrate that SPAN outperforms existing first-order and second-order optimization methods in terms of the convergence wall-clock time. Furthermore, we provide a theoretical analysis of the per-iteration complexity, the approximation error, and the convergence rate. Both the theoretical analysis and experimental results show that our proposed method achieves a better trade-off between the convergence rate and the per-iteration efficiency.

Cite

Text

Huang et al. "SPAN: A Stochastic Projected Approximate Newton Method." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I02.5511

Markdown

[Huang et al. "SPAN: A Stochastic Projected Approximate Newton Method." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/huang2020aaai-span/) doi:10.1609/AAAI.V34I02.5511

BibTeX

@inproceedings{huang2020aaai-span,
  title     = {{SPAN: A Stochastic Projected Approximate Newton Method}},
  author    = {Huang, Xunpeng and Liang, Xianfeng and Liu, Zhengyang and Li, Lei and Yu, Yue and Li, Yitan},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2020},
  pages     = {1520-1527},
  doi       = {10.1609/AAAI.V34I02.5511},
  url       = {https://mlanthology.org/aaai/2020/huang2020aaai-span/}
}