Shallow vs. Deep Sum-Product Networks
Abstract
We investigate the representational power of sum-product networks (computation networks analogous to neural networks, but whose individual units compute either products or weighted sums), through a theoretical analysis that compares deep (multiple hidden layers) vs. shallow (one hidden layer) architectures. We prove there exist families of functions that can be represented much more efficiently with a deep network than with a shallow one, i.e. with substantially fewer hidden units. Such results were not available until now, and contribute to motivate recent research involving learning of deep sum-product networks, and more generally motivate research in Deep Learning.
Cite
Text
Delalleau and Bengio. "Shallow vs. Deep Sum-Product Networks." Neural Information Processing Systems, 2011.Markdown
[Delalleau and Bengio. "Shallow vs. Deep Sum-Product Networks." Neural Information Processing Systems, 2011.](https://mlanthology.org/neurips/2011/delalleau2011neurips-shallow/)BibTeX
@inproceedings{delalleau2011neurips-shallow,
title = {{Shallow vs. Deep Sum-Product Networks}},
author = {Delalleau, Olivier and Bengio, Yoshua},
booktitle = {Neural Information Processing Systems},
year = {2011},
pages = {666-674},
url = {https://mlanthology.org/neurips/2011/delalleau2011neurips-shallow/}
}