Convolutional Gaussian Processes
Abstract
We present a practical way of introducing convolutional structure into Gaussian processes, making them more suited to high-dimensional inputs like images. The main contribution of our work is the construction of an inter-domain inducing point approximation that is well-tailored to the convolutional kernel. This allows us to gain the generalisation benefit of a convolutional kernel, together with fast but accurate posterior inference. We investigate several variations of the convolutional kernel, and apply it to MNIST and CIFAR-10, where we obtain significant improvements over existing Gaussian process models. We also show how the marginal likelihood can be used to find an optimal weighting between convolutional and RBF kernels to further improve performance. This illustration of the usefulness of the marginal likelihood may help automate discovering architectures in larger models.
Cite
Text
van der Wilk et al. "Convolutional Gaussian Processes." Neural Information Processing Systems, 2017.Markdown
[van der Wilk et al. "Convolutional Gaussian Processes." Neural Information Processing Systems, 2017.](https://mlanthology.org/neurips/2017/vanderwilk2017neurips-convolutional/)BibTeX
@inproceedings{vanderwilk2017neurips-convolutional,
title = {{Convolutional Gaussian Processes}},
author = {van der Wilk, Mark and Rasmussen, Carl Edward and Hensman, James},
booktitle = {Neural Information Processing Systems},
year = {2017},
pages = {2849-2858},
url = {https://mlanthology.org/neurips/2017/vanderwilk2017neurips-convolutional/}
}