Gated SoftMax Classification

Abstract

We describe a log-bilinear" model that computes class probabilities by combining an input vector multiplicatively with a vector of binary latent variables. Even though the latent variables can take on exponentially many possible combinations of values, we can efficiently compute the exact probability of each class by marginalizing over the latent variables. This makes it possible to get the exact gradient of the log likelihood. The bilinear score-functions are defined using a three-dimensional weight tensor, and we show that factorizing this tensor allows the model to encode invariances inherent in a task by learning a dictionary of invariant basis functions. Experiments on a set of benchmark problems show that this fully probabilistic model can achieve classification performance that is competitive with (kernel) SVMs, backpropagation, and deep belief nets."

Cite

Text

Memisevic et al. "Gated SoftMax Classification." Neural Information Processing Systems, 2010.

Markdown

[Memisevic et al. "Gated SoftMax Classification." Neural Information Processing Systems, 2010.](https://mlanthology.org/neurips/2010/memisevic2010neurips-gated/)

BibTeX

@inproceedings{memisevic2010neurips-gated,
  title     = {{Gated SoftMax Classification}},
  author    = {Memisevic, Roland and Zach, Christopher and Pollefeys, Marc and Hinton, Geoffrey E.},
  booktitle = {Neural Information Processing Systems},
  year      = {2010},
  pages     = {1603-1611},
  url       = {https://mlanthology.org/neurips/2010/memisevic2010neurips-gated/}
}