Semi-Parametric Inducing Point Networks and Neural Processes

Abstract

We introduce semi-parametric inducing point networks (SPIN), a general-purpose architecture that can query the training set at inference time in a compute-efficient manner. Semi-parametric architectures are typically more compact than parametric models, but their computational complexity is often quadratic. In contrast, SPIN attains linear complexity via a cross-attention mechanism between datapoints inspired by inducing point methods. Querying large training sets can be particularly useful in meta-learning, as it unlocks additional training signal, but often exceeds the scaling limits of existing models. We use SPIN as the basis of the Inducing Point Neural Process, a probabilistic model which supports large contexts in meta-learning and achieves high accuracy where existing models fail. In our experiments, SPIN reduces memory requirements, improves accuracy across a range of meta-learning tasks, and improves state-of-the-art performance on an important practical problem, genotype imputation.

Cite

Text

Rastogi et al. "Semi-Parametric Inducing Point Networks and Neural Processes." International Conference on Learning Representations, 2023.

Markdown

[Rastogi et al. "Semi-Parametric Inducing Point Networks and Neural Processes." International Conference on Learning Representations, 2023.](https://mlanthology.org/iclr/2023/rastogi2023iclr-semiparametric/)

BibTeX

@inproceedings{rastogi2023iclr-semiparametric,
  title     = {{Semi-Parametric Inducing Point Networks and Neural Processes}},
  author    = {Rastogi, Richa and Schiff, Yair and Hacohen, Alon and Li, Zhaozhi and Lee, Ian and Deng, Yuntian and Sabuncu, Mert R. and Kuleshov, Volodymyr},
  booktitle = {International Conference on Learning Representations},
  year      = {2023},
  url       = {https://mlanthology.org/iclr/2023/rastogi2023iclr-semiparametric/}
}