Sparse Activity and Sparse Connectivity in Supervised Learning

Abstract

Sparseness is a useful regularizer for learning in a wide range of applications, in particular in neural networks. This paper proposes a model targeted at classification tasks, where sparse activity and sparse connectivity are used to enhance classification capabilities. The tool for achieving this is a sparseness-enforcing projection operator which finds the closest vector with a pre-defined sparseness for any given vector. In the theoretical part of this paper, a comprehensive theory for such a projection is developed. In conclusion, it is shown that the projection is differentiable almost everywhere and can thus be implemented as a smooth neuronal transfer function. The entire model can hence be tuned end-to-end using gradient-based methods. Experiments on the MNIST database of handwritten digits show that classification performance can be boosted by sparse activity or sparse connectivity. With a combination of both, performance can be significantly better compared to classical non-sparse approaches.

Cite

Text

Thom and Palm. "Sparse Activity and Sparse Connectivity in Supervised Learning." Journal of Machine Learning Research, 2013.

Markdown

[Thom and Palm. "Sparse Activity and Sparse Connectivity in Supervised Learning." Journal of Machine Learning Research, 2013.](https://mlanthology.org/jmlr/2013/thom2013jmlr-sparse/)

BibTeX

@article{thom2013jmlr-sparse,
  title     = {{Sparse Activity and Sparse Connectivity in Supervised Learning}},
  author    = {Thom, Markus and Palm, Günther},
  journal   = {Journal of Machine Learning Research},
  year      = {2013},
  pages     = {1091-1143},
  volume    = {14},
  url       = {https://mlanthology.org/jmlr/2013/thom2013jmlr-sparse/}
}