Exploiting Inferential Structure in Neural Processes

Abstract

Neural Processes (NPs) are appealing due to their ability to perform fast adaptation based on a context set. This set is encoded by a latent variable, which is often assumed to follow a simple distribution. However, in real-word settings, the context set may be drawn from richer distributions having multiple modes, heavy tails, etc. In this work, we provide a framework that allows NPs’ latent variable to be given a rich prior defined by a graphical model. These distributional assumptions directly translate into an appropriate aggregation strategy for the context set. Moreover, we describe a message-passing procedure that still allows for end-to-end optimization with stochastic gradients. We demonstrate the generality of our framework by using mixture and Student-t assumptions that yield improvements in function modelling and test-time robustness.

Cite

Text

Tailor et al. "Exploiting Inferential Structure in Neural Processes." Uncertainty in Artificial Intelligence, 2023.

Markdown

[Tailor et al. "Exploiting Inferential Structure in Neural Processes." Uncertainty in Artificial Intelligence, 2023.](https://mlanthology.org/uai/2023/tailor2023uai-exploiting/)

BibTeX

@inproceedings{tailor2023uai-exploiting,
  title     = {{Exploiting Inferential Structure in Neural Processes}},
  author    = {Tailor, Dharmesh and Khan, Mohammad Emtiyaz and Nalisnick, Eric},
  booktitle = {Uncertainty in Artificial Intelligence},
  year      = {2023},
  pages     = {2089-2098},
  volume    = {216},
  url       = {https://mlanthology.org/uai/2023/tailor2023uai-exploiting/}
}