Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs
Abstract
Sparse high-dimensional data vectors are common in many application domains where a very large number of rarely non-zero features can be devised. Unfortunately, this creates a computational bottleneck for unsupervised feature learning algorithms such as those based on auto-encoders and RBMs, because they involve a reconstruction step where the whole input vector is predicted from the current feature values. An algorithm was recently developed to successfully handle the case of auto-encoders, based on an importance sampling scheme stochastically selecting which input elements to actually reconstruct during training for each particular example. To generalize this idea to RBMs, we propose a stochastic ratio-matching algorithm that inherits all the computational advantages and unbiasedness of the importance sampling scheme. We show that stochastic ratio matching is a good estimator, allowing the approach to beat the state-of-the-art on two bag-of-word text classification benchmarks (20 Newsgroups and RCV1), while keeping computational cost linear in the number of non-zeros.
Cite
Text
Dauphin and Bengio. "Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs." Neural Information Processing Systems, 2013.Markdown
[Dauphin and Bengio. "Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/dauphin2013neurips-stochastic/)BibTeX
@inproceedings{dauphin2013neurips-stochastic,
title = {{Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs}},
author = {Dauphin, Yann and Bengio, Yoshua},
booktitle = {Neural Information Processing Systems},
year = {2013},
pages = {1340-1348},
url = {https://mlanthology.org/neurips/2013/dauphin2013neurips-stochastic/}
}