From Boltzmann Machines to Neural Networks and Back Again
Abstract
Graphical models are powerful tools for modeling high-dimensional data, but learning graphical models in the presence of latent variables is well-known to be difficult. In this work we give new results for learning Restricted Boltzmann Machines, probably the most well-studied class of latent variable models. Our results are based on new connections to learning two-layer neural networks under $\ell_{\infty}$ bounded input; for both problems, we give nearly optimal results under the conjectured hardness of sparse parity with noise. Using the connection between RBMs and feedforward networks, we also initiate the theoretical study of {\em supervised RBMs} \citep{hinton2012practical}, a version of neural-network learning that couples distributional assumptions induced from the underlying graphical model with the architecture of the unknown function class. We then give an algorithm for learning a natural class of supervised RBMs with better runtime than what is possible for its related class of networks without distributional assumptions.
Cite
Text
Goel et al. "From Boltzmann Machines to Neural Networks and Back Again." Neural Information Processing Systems, 2020.Markdown
[Goel et al. "From Boltzmann Machines to Neural Networks and Back Again." Neural Information Processing Systems, 2020.](https://mlanthology.org/neurips/2020/goel2020neurips-boltzmann/)BibTeX
@inproceedings{goel2020neurips-boltzmann,
title = {{From Boltzmann Machines to Neural Networks and Back Again}},
author = {Goel, Surbhi and Klivans, Adam and Koehler, Frederic},
booktitle = {Neural Information Processing Systems},
year = {2020},
url = {https://mlanthology.org/neurips/2020/goel2020neurips-boltzmann/}
}