Exploration of Autoregressive Models for In-Context Learning on Tabular Data

Abstract

We explore different auto-regressive model architectures for in-context learning on tabular datasets trained in a similar manner to TabPFN. Namely, we compare transformer based models with a structured state-space model architecture (Mamba) and a hybrid architecture (Jamba), mixing transformer and Mamba layers. We find that auto-regressive transformer models perform similarly to the original TabPFN transformer architectures, albeit at the cost of a doubled context length. Mamba performs worse than similar sized transformer models, while hybrid models show promise in harnessing some advantages of state-space models such as supporting long input context length and fast inference.

Cite

Text

Baur and Kim. "Exploration of Autoregressive Models for In-Context Learning on Tabular Data." NeurIPS 2024 Workshops: TRL, 2024.

Markdown

[Baur and Kim. "Exploration of Autoregressive Models for In-Context Learning on Tabular Data." NeurIPS 2024 Workshops: TRL, 2024.](https://mlanthology.org/neuripsw/2024/baur2024neuripsw-exploration/)

BibTeX

@inproceedings{baur2024neuripsw-exploration,
  title     = {{Exploration of Autoregressive Models for In-Context Learning on Tabular Data}},
  author    = {Baur, Stefan K. and Kim, Sohyeong},
  booktitle = {NeurIPS 2024 Workshops: TRL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/baur2024neuripsw-exploration/}
}