Assumed Density Filtering Methods for Learning Bayesian Neural Networks

Abstract

Buoyed by the success of deep multilayer neural networks, there is renewed interest in scalable learning of Bayesian neural networks. Here, we study algorithms that utilize recent advances in Bayesian inference to efficiently learn distributions over network weights. In particular, we focus on recently proposed assumed density filtering based methods for learning Bayesian neural networks -- Expectation and Probabilistic backpropagation. Apart from scaling to large datasets, these techniques seamlessly deal with non-differentiable activation functions and provide parameter (learning rate, momentum) free learning. In this paper, we first rigorously compare the two algorithms and in the process develop several extensions, including a version of EBP for continuous regression problems and a PBP variant for binary classification. Next, we extend both algorithms to deal with multiclass classification and count regression problems. On a variety of diverse real world benchmarks, we find our extensions to be effective, achieving results competitive with the state-of-the-art.

Cite

Text

Ghosh et al. "Assumed Density Filtering Methods for Learning Bayesian Neural Networks." AAAI Conference on Artificial Intelligence, 2016. doi:10.1609/AAAI.V30I1.10296

Markdown

[Ghosh et al. "Assumed Density Filtering Methods for Learning Bayesian Neural Networks." AAAI Conference on Artificial Intelligence, 2016.](https://mlanthology.org/aaai/2016/ghosh2016aaai-assumed/) doi:10.1609/AAAI.V30I1.10296

BibTeX

@inproceedings{ghosh2016aaai-assumed,
  title     = {{Assumed Density Filtering Methods for Learning Bayesian Neural Networks}},
  author    = {Ghosh, Soumya and Fave, Francesco Maria Delle and Yedidia, Jonathan S.},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2016},
  pages     = {1589-1595},
  doi       = {10.1609/AAAI.V30I1.10296},
  url       = {https://mlanthology.org/aaai/2016/ghosh2016aaai-assumed/}
}