Emergence of Implicit Filter Sparsity in Convolutional Neural Networks

Abstract

We show implicit filter level sparsity manifests in convolutional neural networks (CNNs) which employ Batch Normalization and ReLU activation, and are trained using adaptive gradient descent techniques with L2 regularization or weight decay. Through an extensive empirical study (Anonymous, 2019) we hypothesize the mechanism be hind the sparsification process. We find that the interplay of various phenomena influences the strength of L2 and weight decay regularizers, leading the supposedly non sparsity inducing regularizers to induce filter sparsity. In this workshop article we summarize some of our key findings and experiments, and present additional results on modern network architectures such as ResNet-50.

Cite

Text

Mehta et al. "Emergence of Implicit Filter Sparsity in Convolutional Neural Networks." ICML 2019 Workshops: Deep_Phenomena, 2019.

Markdown

[Mehta et al. "Emergence of Implicit Filter Sparsity in Convolutional Neural Networks." ICML 2019 Workshops: Deep_Phenomena, 2019.](https://mlanthology.org/icmlw/2019/mehta2019icmlw-emergence/)

BibTeX

@inproceedings{mehta2019icmlw-emergence,
  title     = {{Emergence of Implicit Filter Sparsity in Convolutional Neural Networks}},
  author    = {Mehta, Dushyant and Kim, Kwang In and Theobalt, Christian},
  booktitle = {ICML 2019 Workshops: Deep_Phenomena},
  year      = {2019},
  url       = {https://mlanthology.org/icmlw/2019/mehta2019icmlw-emergence/}
}