Training Compressed Fully-Connected Networks with a Density-Diversity Penalty
Abstract
Deep models have achieved great success on a variety of challenging tasks. How- ever, the models that achieve great performance often have an enormous number of parameters, leading to correspondingly great demands on both computational and memory resources, especially for fully-connected layers. In this work, we propose a new “density-diversity penalty” regularizer that can be applied to fully-connected layers of neural networks during training. We show that using this regularizer results in significantly fewer parameters (i.e., high sparsity), and also significantly fewer distinct values (i.e., low diversity), so that the trained weight matrices can be highly compressed without any appreciable loss in performance. The resulting trained models can hence reside on computational platforms (e.g., portables, Internet-of-Things devices) where it otherwise would be prohibitive.
Cite
Text
Wang et al. "Training Compressed Fully-Connected Networks with a Density-Diversity Penalty." International Conference on Learning Representations, 2017.Markdown
[Wang et al. "Training Compressed Fully-Connected Networks with a Density-Diversity Penalty." International Conference on Learning Representations, 2017.](https://mlanthology.org/iclr/2017/wang2017iclr-training/)BibTeX
@inproceedings{wang2017iclr-training,
title = {{Training Compressed Fully-Connected Networks with a Density-Diversity Penalty}},
author = {Wang, Shengjie and Cai, Haoran and Bilmes, Jeff A. and Noble, William S.},
booktitle = {International Conference on Learning Representations},
year = {2017},
url = {https://mlanthology.org/iclr/2017/wang2017iclr-training/}
}