SPAN: A Stochastic Projected Approximate Newton Method
Abstract
Second-order optimization methods have desirable convergence properties. However, the exact Newton method requires expensive computation for the Hessian and its inverse. In this paper, we propose SPAN, a novel approximate and fast Newton method. SPAN computes the inverse of the Hessian matrix via low-rank approximation and stochastic Hessian-vector products. Our experiments on multiple benchmark datasets demonstrate that SPAN outperforms existing first-order and second-order optimization methods in terms of the convergence wall-clock time. Furthermore, we provide a theoretical analysis of the per-iteration complexity, the approximation error, and the convergence rate. Both the theoretical analysis and experimental results show that our proposed method achieves a better trade-off between the convergence rate and the per-iteration efficiency.
Cite
Text
Huang et al. "SPAN: A Stochastic Projected Approximate Newton Method." AAAI Conference on Artificial Intelligence, 2020. doi:10.1609/AAAI.V34I02.5511Markdown
[Huang et al. "SPAN: A Stochastic Projected Approximate Newton Method." AAAI Conference on Artificial Intelligence, 2020.](https://mlanthology.org/aaai/2020/huang2020aaai-span/) doi:10.1609/AAAI.V34I02.5511BibTeX
@inproceedings{huang2020aaai-span,
title = {{SPAN: A Stochastic Projected Approximate Newton Method}},
author = {Huang, Xunpeng and Liang, Xianfeng and Liu, Zhengyang and Li, Lei and Yu, Yue and Li, Yitan},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2020},
pages = {1520-1527},
doi = {10.1609/AAAI.V34I02.5511},
url = {https://mlanthology.org/aaai/2020/huang2020aaai-span/}
}