An Empirical Study of the $(l_0, L_1)$-Smoothness Condition
Abstract
The $(L_0,L_1)$-smoothness condition was introduced by Zhang-He-Sra-Jadbabai in 2020, who both proved convergence bounds under this assumption and provided empirical evidence it is satisfied in deep learning. Since then, many groups have proven convergence guarantees for functions which satisfy this condition, motivated by the expectation that loss functions arising in deep learning satisfy it. In this paper we provide further empirical study of this condition in the setting of feedforward neural networks of depth at least 2, with $L2$ or cross entropy loss. The results suggest that the $(L_0,L_1)$-smoothness condition is not satisfied in this setting.
Cite
Text
Cooper. "An Empirical Study of the $(l_0, L_1)$-Smoothness Condition." NeurIPS 2024 Workshops: M3L, 2024.Markdown
[Cooper. "An Empirical Study of the $(l_0, L_1)$-Smoothness Condition." NeurIPS 2024 Workshops: M3L, 2024.](https://mlanthology.org/neuripsw/2024/cooper2024neuripsw-empirical/)BibTeX
@inproceedings{cooper2024neuripsw-empirical,
title = {{An Empirical Study of the $(l_0, L_1)$-Smoothness Condition}},
author = {Cooper, Y},
booktitle = {NeurIPS 2024 Workshops: M3L},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/cooper2024neuripsw-empirical/}
}