Subspace Inference for Bayesian Deep Learning
Abstract
Bayesian inference was once a gold standard for learning with neural networks, providing accurate full predictive distributions and well calibrated uncertainty. However, scaling Bayesian inference techniques to deep neural networks is challenging due to the high dimensionality of the parameter space. In this paper, we construct low-dimensional subspaces of parameter space, such as the first principal components of the stochastic gradient descent (SGD) trajectory, which contain diverse sets of high performing models. In these subspaces, we are able to apply elliptical slice sampling and variational inference, which struggle in the full parameter space. We show that Bayesian model averaging over the induced posterior in these subspaces produces accurate predictions and well-calibrated predictive uncertainty for both regression and image classification.
Cite
Text
Izmailov et al. "Subspace Inference for Bayesian Deep Learning." Uncertainty in Artificial Intelligence, 2019.Markdown
[Izmailov et al. "Subspace Inference for Bayesian Deep Learning." Uncertainty in Artificial Intelligence, 2019.](https://mlanthology.org/uai/2019/izmailov2019uai-subspace/)BibTeX
@inproceedings{izmailov2019uai-subspace,
title = {{Subspace Inference for Bayesian Deep Learning}},
author = {Izmailov, Pavel and Maddox, Wesley J. and Kirichenko, Polina and Garipov, Timur and Vetrov, Dmitry and Wilson, Andrew Gordon},
booktitle = {Uncertainty in Artificial Intelligence},
year = {2019},
pages = {1169-1179},
volume = {115},
url = {https://mlanthology.org/uai/2019/izmailov2019uai-subspace/}
}