Zeroth-Order Negative Curvature Finding: Escaping Saddle Points Without Gradients
Abstract
We consider escaping saddle points of nonconvex problems where only the function evaluations can be accessed. Although a variety of works have been proposed, the majority of them require either second or first-order information, and only a few of them have exploited zeroth-order methods, particularly the technique of negative curvature finding with zeroth-order methods which has been proven to be the most efficient method for escaping saddle points. To fill this gap, in this paper, we propose two zeroth-order negative curvature finding frameworks that can replace Hessian-vector product computations without increasing the iteration complexity. We apply the proposed frameworks to ZO-GD, ZO-SGD, ZO-SCSG, ZO-SPIDER and prove that these ZO algorithms can converge to $(\epsilon,\delta)$-approximate second-order stationary points with less query complexity compared with prior zeroth-order works for finding local minima.
Cite
Text
Zhang et al. "Zeroth-Order Negative Curvature Finding: Escaping Saddle Points Without Gradients." Neural Information Processing Systems, 2022.Markdown
[Zhang et al. "Zeroth-Order Negative Curvature Finding: Escaping Saddle Points Without Gradients." Neural Information Processing Systems, 2022.](https://mlanthology.org/neurips/2022/zhang2022neurips-zerothorder/)BibTeX
@inproceedings{zhang2022neurips-zerothorder,
title = {{Zeroth-Order Negative Curvature Finding: Escaping Saddle Points Without Gradients}},
author = {Zhang, Hualin and Xiong, Huan and Gu, Bin},
booktitle = {Neural Information Processing Systems},
year = {2022},
url = {https://mlanthology.org/neurips/2022/zhang2022neurips-zerothorder/}
}