Shallow vs. Deep Sum-Product Networks

Abstract

We investigate the representational power of sum-product networks (computation networks analogous to neural networks, but whose individual units compute either products or weighted sums), through a theoretical analysis that compares deep (multiple hidden layers) vs. shallow (one hidden layer) architectures. We prove there exist families of functions that can be represented much more efficiently with a deep network than with a shallow one, i.e. with substantially fewer hidden units. Such results were not available until now, and contribute to motivate recent research involving learning of deep sum-product networks, and more generally motivate research in Deep Learning.

Cite

Text

Delalleau and Bengio. "Shallow vs. Deep Sum-Product Networks." Neural Information Processing Systems, 2011.

Markdown

[Delalleau and Bengio. "Shallow vs. Deep Sum-Product Networks." Neural Information Processing Systems, 2011.](https://mlanthology.org/neurips/2011/delalleau2011neurips-shallow/)

BibTeX

@inproceedings{delalleau2011neurips-shallow,
  title     = {{Shallow vs. Deep Sum-Product Networks}},
  author    = {Delalleau, Olivier and Bengio, Yoshua},
  booktitle = {Neural Information Processing Systems},
  year      = {2011},
  pages     = {666-674},
  url       = {https://mlanthology.org/neurips/2011/delalleau2011neurips-shallow/}
}