On the Abilities of Mathematical Extrapolation with Implicit Models

Abstract

Deep neural networks excel on a variety of different tasks, often surpassing human intelligence. However, when presented with out-of-distribution data, these models tend to break down even on the simplest tasks. In this paper, we compare the robustness of implicitly-defined and classical deep learning models on a series of mathematical extrapolation tasks, where the models are tested with out-of-distribution samples during inference time. Throughout our experiments, implicit models greatly outperform classical deep learning networks that overfit the training distribution. We present implicit models as a safer deep learning framework for generalization due to their flexible and selective structure. Implicit models, with potentially unlimited depth, not only adapt well to out-of-distribution data but also understand the underlying structure of inputs much better.

Cite

Text

Decugis et al. "On the Abilities of Mathematical Extrapolation with Implicit Models." NeurIPS 2022 Workshops: MLSW, 2022.

Markdown

[Decugis et al. "On the Abilities of Mathematical Extrapolation with Implicit Models." NeurIPS 2022 Workshops: MLSW, 2022.](https://mlanthology.org/neuripsw/2022/decugis2022neuripsw-abilities-a/)

BibTeX

@inproceedings{decugis2022neuripsw-abilities-a,
  title     = {{On the Abilities of Mathematical Extrapolation with Implicit Models}},
  author    = {Decugis, Juliette and Tsai, Alicia Y. and Ganesh, Ashwin and Emerling, Max and El Ghaoui, Laurent},
  booktitle = {NeurIPS 2022 Workshops: MLSW},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/decugis2022neuripsw-abilities-a/}
}