Bayesian Neural Network Priors Revisited
Abstract
Isotropic Gaussian priors are the de facto standard for modern Bayesian neural network inference. However, such simplistic priors are unlikely to either accurately reflect our true beliefs about the weight distributions, or to give optimal performance. We study summary statistics of (convolutional) neural network weights in networks trained using SGD. We find that in certain circumstances, these networks have heavy-tailed weight distributions, while convolutional neural network weights often display strong spatial correlations. Building these observations into the respective priors, we get improved performance on MNIST classification. Remarkably, we find that using a more accurate prior partially mitigates the cold posterior effect, by improving performance at high temperatures corresponding to exact Bayesian inference, while having less of an effect at small temperatures.
Cite
Text
Fortuin et al. "Bayesian Neural Network Priors Revisited." NeurIPS 2020 Workshops: ICBINB, 2020.Markdown
[Fortuin et al. "Bayesian Neural Network Priors Revisited." NeurIPS 2020 Workshops: ICBINB, 2020.](https://mlanthology.org/neuripsw/2020/fortuin2020neuripsw-bayesian/)BibTeX
@inproceedings{fortuin2020neuripsw-bayesian,
title = {{Bayesian Neural Network Priors Revisited}},
author = {Fortuin, Vincent and Garriga-Alonso, Adrià and Wenzel, Florian and Ratsch, Gunnar and Turner, Richard E and van der Wilk, Mark and Aitchison, Laurence},
booktitle = {NeurIPS 2020 Workshops: ICBINB},
year = {2020},
url = {https://mlanthology.org/neuripsw/2020/fortuin2020neuripsw-bayesian/}
}