Representation Costs of Linear Neural Networks: Analysis and Design
Abstract
For different parameterizations (mappings from parameters to predictors), we study the regularization cost in predictor space induced by $l_2$ regularization on the parameters (weights). We focus on linear neural networks as parameterizations of linear predictors. We identify the representation cost of certain sparse linear ConvNets and residual networks. In order to get a better understanding of how the architecture and parameterization affect the representation cost, we also study the reverse problem, identifying which regularizers on linear predictors (e.g., $l_p$ norms, group norms, the $k$-support-norm, elastic net) can be the representation cost induced by simple $l_2$ regularization, and designing the parameterizations that do so.
Cite
Text
Dai et al. "Representation Costs of Linear Neural Networks: Analysis and Design." Neural Information Processing Systems, 2021.Markdown
[Dai et al. "Representation Costs of Linear Neural Networks: Analysis and Design." Neural Information Processing Systems, 2021.](https://mlanthology.org/neurips/2021/dai2021neurips-representation/)BibTeX
@inproceedings{dai2021neurips-representation,
title = {{Representation Costs of Linear Neural Networks: Analysis and Design}},
author = {Dai, Zhen and Karzand, Mina and Srebro, Nathan},
booktitle = {Neural Information Processing Systems},
year = {2021},
url = {https://mlanthology.org/neurips/2021/dai2021neurips-representation/}
}