Lower Bounds for Smooth Nonconvex Finite-Sum Optimization
Abstract
Smooth finite-sum optimization has been widely studied in both convex and nonconvex settings. However, existing lower bounds for finite-sum optimization are mostly limited to the setting where each component function is (strongly) convex, while the lower bounds for nonconvex finite-sum optimization remain largely unsolved. In this paper, we study the lower bounds for smooth nonconvex finite-sum optimization, where the objective function is the average of $n$ nonconvex component functions. We prove tight lower bounds for the complexity of finding $\epsilon$-suboptimal point and $\epsilon$-approximate stationary point in different settings, for a wide regime of the smallest eigenvalue of the Hessian of the objective function (or each component function). Given our lower bounds, we can show that existing algorithms including KatyushaX \citep{allen2018katyushax}, Natasha \citep{allen2017natasha} and StagewiseKatyusha \citep{yang2018does} have achieved optimal Incremental First-order Oracle (IFO) complexity (i.e., number of IFO calls) up to logarithm factors for nonconvex finite-sum optimization. We also point out potential ways to further improve these complexity results, in terms of making stronger assumptions or by a different convergence analysis.
Cite
Text
Zhou and Gu. "Lower Bounds for Smooth Nonconvex Finite-Sum Optimization." International Conference on Machine Learning, 2019.Markdown
[Zhou and Gu. "Lower Bounds for Smooth Nonconvex Finite-Sum Optimization." International Conference on Machine Learning, 2019.](https://mlanthology.org/icml/2019/zhou2019icml-lower/)BibTeX
@inproceedings{zhou2019icml-lower,
title = {{Lower Bounds for Smooth Nonconvex Finite-Sum Optimization}},
author = {Zhou, Dongruo and Gu, Quanquan},
booktitle = {International Conference on Machine Learning},
year = {2019},
pages = {7574-7583},
volume = {97},
url = {https://mlanthology.org/icml/2019/zhou2019icml-lower/}
}