Structured In-Context Task Representations
Abstract
Representation learning has been central to deep learning’s evolution. While interpretable structures have been observed in pre-trained models’ representations, an important question arises: Do networks develop such interpretable structures during in-context learning? Using synthetic sequence data derived from underlying geometrically structured graphs (e.g., grids, rings), we provide affirmative evidence that language models develop internal representations mirroring these geometric structures during in-context learning. Furthermore, we demonstrate how in-context examples can override semantic priors by constructing a representation in dimensions other than the one used by the prior. Overall, our study demonstrates that models can form meaningful representations solely from in-context exemplars.
Cite
Text
Park et al. "Structured In-Context Task Representations." NeurIPS 2024 Workshops: NeurReps, 2024.Markdown
[Park et al. "Structured In-Context Task Representations." NeurIPS 2024 Workshops: NeurReps, 2024.](https://mlanthology.org/neuripsw/2024/park2024neuripsw-structured/)BibTeX
@inproceedings{park2024neuripsw-structured,
title = {{Structured In-Context Task Representations}},
author = {Park, Core Francisco and Lee, Andrew and Lubana, Ekdeep Singh and Nishi, Kento and Okawa, Maya and Tanaka, Hidenori},
booktitle = {NeurIPS 2024 Workshops: NeurReps},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/park2024neuripsw-structured/}
}