Implicit Regularization and Convergence for Weight Normalization
Abstract
Normalization methods such as batch, weight, instance, and layer normalization are commonly used in modern machine learning. Here, we study the weight normalization (WN) method \cite{salimans2016weight} and a variant called reparametrized projected gradient descent (rPGD) for overparametrized least squares regression and some more general loss functions. WN and rPGD reparametrize the weights with a scale $g$ and a unit vector such that the objective function becomes \emph{non-convex}. We show that this non-convex formulation has beneficial regularization effects compared to gradient descent on the original objective. These methods adaptively regularize the weights and \emph{converge linearly} close to the minimum $\ell_2$ norm solution even for initializations far from zero. For certain two-phase variants, they can converge to the min norm solution. This is different from the behavior of gradient descent, which only converges to the min norm solution when started at zero, and thus more sensitive to initialization.
Cite
Text
Wu et al. "Implicit Regularization and Convergence for Weight Normalization." Neural Information Processing Systems, 2020.Markdown
[Wu et al. "Implicit Regularization and Convergence for Weight Normalization." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/wu2020neurips-implicit/)BibTeX
@inproceedings{wu2020neurips-implicit,
title = {{Implicit Regularization and Convergence for Weight Normalization}},
author = {Wu, Xiaoxia and Dobriban, Edgar and Ren, Tongzheng and Wu, Shanshan and Li, Zhiyuan and Gunasekar, Suriya and Ward, Rachel and Liu, Qiang},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/wu2020neurips-implicit/}
}