Amortized Variational Deep Kernel Learning

Abstract

Deep kernel learning (DKL) marries the uncertainty quantification of Gaussian processes (GPs) and the representational power of deep neural networks. However, training DKL is challenging and often leads to overfitting. Most notably, DKL often learns “non-local” kernels — incurring spurious correlations. To remedy this issue, we propose using amortized inducing points and a parameter-sharing scheme, which ties together the amortization and DKL networks. This design imposes an explicit dependency between the ELBO’s model fit and capacity terms. In turn, this prevents the former from dominating the optimization procedure and incurring the aforementioned spurious correlations. Extensive experiments show that our resulting method, amortized varitional DKL (AVDKL), i) consistently outperforms DKL and standard GPs for tabular data; ii) achieves significantly higher accuracy than DKL in node classification tasks; and iii) leads to substantially better accuracy and negative log-likelihood than DKL on CIFAR100.

Cite

Text

Matias et al. "Amortized Variational Deep Kernel Learning." International Conference on Machine Learning, 2024.

Markdown

[Matias et al. "Amortized Variational Deep Kernel Learning." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/matias2024icml-amortized/)

BibTeX

@inproceedings{matias2024icml-amortized,
  title     = {{Amortized Variational Deep Kernel Learning}},
  author    = {Matias, Alan L. S. and Mattos, César Lincoln and Gomes, João Paulo Pordeus and Mesquita, Diego},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {35063-35078},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/matias2024icml-amortized/}
}