Residual Sum-Product Networks
Abstract
Tractable yet expressive density estimators are a key building block of probabilistic machine learning. While sum-product networks (SPNs) offer attractive inference capabilities, obtaining structures large enough to fit complex, high-dimensional data has proven challenging. In this paper, we present a residual learning approach to ease the learning of SPNs, which are deeper and wider than those used previously. The main trick is to ensemble SPNs by explicitly reformulating sum nodes as residual functions. This adds references to substructures across the SPNs at different depths, which in turn helps to improve training. Our experiments demonstrate that the resulting residual SPNs (ResSPNs) are easy to optimize, gain performance from considerably increased depth and width, and are competitive to state of-the-art SPN structure learning approaches. To combat overfitting, we introduce an iterative pruning technique that compacts models and yields better generalization.
Cite
Text
Ventola et al. "Residual Sum-Product Networks." Proceedings of pgm 2020, 2020.Markdown
[Ventola et al. "Residual Sum-Product Networks." Proceedings of pgm 2020, 2020.](https://mlanthology.org/pgm/2020/ventola2020pgm-residual/)BibTeX
@inproceedings{ventola2020pgm-residual,
title = {{Residual Sum-Product Networks}},
author = {Ventola, Fabrizio and Stelzner, Karl and Molina, Alejandro and Kersting, Kristian},
booktitle = {Proceedings of pgm 2020},
year = {2020},
pages = {545-556},
volume = {138},
url = {https://mlanthology.org/pgm/2020/ventola2020pgm-residual/}
}