Tabby: Tabular Adaptation for Language Models
Abstract
While the quality of synthetic text data has greatly improved in recent years, thanks to specialized architectures such as large language models (LLMs), tabular data has received relatively less attention. To address this disparity, we present Tabby: a modification to the LLM architecture that enables its use for tabular dataset synthesis. Tabby consists of a novel adaptation of Gated Mixture-of-Expert layers, allowing each data column to be modeled by dedicated parameters within the Transformer multi-layer perceptrons or language modeling head. Applying Tabby to Distilled GPT-2 improves synthetic data quality (measured by machine learning efficacy) by up to 2.7% compared to previous tabular dataset synthesis methods, achieving performance near or equal to that of real data.
Cite
Text
Cromp et al. "Tabby: Tabular Adaptation for Language Models." NeurIPS 2024 Workshops: TRL, 2024.Markdown
[Cromp et al. "Tabby: Tabular Adaptation for Language Models." NeurIPS 2024 Workshops: TRL, 2024.](https://mlanthology.org/neuripsw/2024/cromp2024neuripsw-tabby/)BibTeX
@inproceedings{cromp2024neuripsw-tabby,
title = {{Tabby: Tabular Adaptation for Language Models}},
author = {Cromp, Sonia and Gnvv, Satya Sai Srinath Namburi and Cao, Catherine and Alkhudhayri, Mohammed and Guo, Samuel and Roberts, Nicholas and Sala, Frederic},
booktitle = {NeurIPS 2024 Workshops: TRL},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/cromp2024neuripsw-tabby/}
}