Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs

Abstract

Sparse high-dimensional data vectors are common in many application domains where a very large number of rarely non-zero features can be devised. Unfortunately, this creates a computational bottleneck for unsupervised feature learning algorithms such as those based on auto-encoders and RBMs, because they involve a reconstruction step where the whole input vector is predicted from the current feature values. An algorithm was recently developed to successfully handle the case of auto-encoders, based on an importance sampling scheme stochastically selecting which input elements to actually reconstruct during training for each particular example. To generalize this idea to RBMs, we propose a stochastic ratio-matching algorithm that inherits all the computational advantages and unbiasedness of the importance sampling scheme. We show that stochastic ratio matching is a good estimator, allowing the approach to beat the state-of-the-art on two bag-of-word text classification benchmarks (20 Newsgroups and RCV1), while keeping computational cost linear in the number of non-zeros.

Cite

Text

Dauphin and Bengio. "Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs." Neural Information Processing Systems, 2013.

Markdown

[Dauphin and Bengio. "Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs." Neural Information Processing Systems, 2013.](https://mlanthology.org/neurips/2013/dauphin2013neurips-stochastic/)

BibTeX

@inproceedings{dauphin2013neurips-stochastic,
  title     = {{Stochastic Ratio Matching of RBMs for Sparse High-Dimensional Inputs}},
  author    = {Dauphin, Yann and Bengio, Yoshua},
  booktitle = {Neural Information Processing Systems},
  year      = {2013},
  pages     = {1340-1348},
  url       = {https://mlanthology.org/neurips/2013/dauphin2013neurips-stochastic/}
}