Exploration of Autoregressive Models for In-Context Learning on Tabular Data
Abstract
We explore different auto-regressive model architectures for in-context learning on tabular datasets trained in a similar manner to TabPFN. Namely, we compare transformer based models with a structured state-space model architecture (Mamba) and a hybrid architecture (Jamba), mixing transformer and Mamba layers. We find that auto-regressive transformer models perform similarly to the original TabPFN transformer architectures, albeit at the cost of a doubled context length. Mamba performs worse than similar sized transformer models, while hybrid models show promise in harnessing some advantages of state-space models such as supporting long input context length and fast inference.
Cite
Text
Baur and Kim. "Exploration of Autoregressive Models for In-Context Learning on Tabular Data." NeurIPS 2024 Workshops: TRL, 2024.Markdown
[Baur and Kim. "Exploration of Autoregressive Models for In-Context Learning on Tabular Data." NeurIPS 2024 Workshops: TRL, 2024.](https://mlanthology.org/neuripsw/2024/baur2024neuripsw-exploration/)BibTeX
@inproceedings{baur2024neuripsw-exploration,
title = {{Exploration of Autoregressive Models for In-Context Learning on Tabular Data}},
author = {Baur, Stefan K. and Kim, Sohyeong},
booktitle = {NeurIPS 2024 Workshops: TRL},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/baur2024neuripsw-exploration/}
}