How Capable Can a Transformer Become? a Study on Synthetic, Interpretable Tasks
Abstract
Transformers trained on huge text corpora exhibit a remarkable set of capabilities. Given the inherent compositional nature of language, one can expect the model to learn to compose these capabilities, potentially yielding a combinatorial explosion of what operations it can perform on an input. Motivated by the above, we aim to assess in this paper "how capable can a transformer become?". In this work, we train Transformer models on a data-generating process that involves compositions of a set of well-defined monolithic capabilities and show that: (1) Transformers generalize to exponentially or even combinatorially many functions not seen in the training data; (2) Transformers that generate the intermediate outputs of the composition are more effective at generalizing to unseen compositions; (3) The training data has a significant impact on the model's ability to compose functions (4) Attention layers in the latter half of the model seem critical to compositionality.
Cite
Text
Ramesh et al. "How Capable Can a Transformer Become? a Study on Synthetic, Interpretable Tasks." NeurIPS 2023 Workshops: R0-FoMo, 2023.Markdown
[Ramesh et al. "How Capable Can a Transformer Become? a Study on Synthetic, Interpretable Tasks." NeurIPS 2023 Workshops: R0-FoMo, 2023.](https://mlanthology.org/neuripsw/2023/ramesh2023neuripsw-capable-a/)BibTeX
@inproceedings{ramesh2023neuripsw-capable-a,
title = {{How Capable Can a Transformer Become? a Study on Synthetic, Interpretable Tasks}},
author = {Ramesh, Rahul and Khona, Mikail and Dick, Robert P. and Tanaka, Hidenori and Lubana, Ekdeep Singh},
booktitle = {NeurIPS 2023 Workshops: R0-FoMo},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/ramesh2023neuripsw-capable-a/}
}