Learning Deep Energy Models
Abstract
Deep generative models with multiple hidden layers have been shown to be able to learn meaningful and compact representations of data. In this work we propose deep energy models, a class of models that use a deep feedforward neural network to model the energy landscape that defines a probabilistic model. We are able to efficiently train all layers of our model at the same time, allowing the lower layers of the model to adapt to the training of the higher layers, producing better generative models. We evaluate the generative performance of our models on natural images and demonstrate that joint training of multiple layers yields qualitative and quantitative improvements over greedy layerwise training. We further generalize our models beyond the commonly used sigmoidal neural networks and show how a deep extension of the product of Student-t distributions model achieves good generative performance. Finally, we introduce a discriminative extension of our model and demonstrate that it outperforms other fully-connected models on object recognition on the NORB dataset.
Cite
Text
Ngiam et al. "Learning Deep Energy Models." International Conference on Machine Learning, 2011.Markdown
[Ngiam et al. "Learning Deep Energy Models." International Conference on Machine Learning, 2011.](https://mlanthology.org/icml/2011/ngiam2011icml-learning/)BibTeX
@inproceedings{ngiam2011icml-learning,
title = {{Learning Deep Energy Models}},
author = {Ngiam, Jiquan and Chen, Zhenghao and Koh, Pang Wei and Ng, Andrew Y.},
booktitle = {International Conference on Machine Learning},
year = {2011},
pages = {1105-1112},
url = {https://mlanthology.org/icml/2011/ngiam2011icml-learning/}
}