Width Provably Matters in Optimization for Deep Linear Neural Networks
Abstract
We prove that for an $L$-layer fully-connected linear neural network, if the width of every hidden layer is $\widetilde{\Omega}\left(L \cdot r \cdot d_{out} \cdot \kappa^3 \right)$, where $r$ and $\kappa$ are the rank and the condition number of the input data, and $d_{out}$ is the output dimension, then gradient descent with Gaussian random initialization converges to a global minimum at a linear rate. The number of iterations to find an $\epsilon$-suboptimal solution is $O(\kappa \log(\frac{1}{\epsilon}))$. Our polynomial upper bound on the total running time for wide deep linear networks and the $\exp\left(\Omega\left(L\right)\right)$ lower bound for narrow deep linear neural networks [Shamir, 2018] together demonstrate that wide layers are necessary for optimizing deep models.
Cite
Text
Du and Hu. "Width Provably Matters in Optimization for Deep Linear Neural Networks." International Conference on Machine Learning, 2019.Markdown
[Du and Hu. "Width Provably Matters in Optimization for Deep Linear Neural Networks." International Conference on Machine Learning, 2019.](https://mlanthology.org/icml/2019/du2019icml-width/)BibTeX
@inproceedings{du2019icml-width,
title = {{Width Provably Matters in Optimization for Deep Linear Neural Networks}},
author = {Du, Simon and Hu, Wei},
booktitle = {International Conference on Machine Learning},
year = {2019},
pages = {1655-1664},
volume = {97},
url = {https://mlanthology.org/icml/2019/du2019icml-width/}
}