How Do Language Models Bind Entities in Context?

Abstract

Language models (LMs) can recall facts mentioned in context, as shown by their performance on reading comprehension tasks. When the context describes facts about more than one entity, the LM has to correctly bind attributes to their corresponding entity. We show, via causal experiments, that LMs' internal activations represent binding information by exhibiting appropriate binding ID vectors at the entity and attribute positions. We further show that binding ID vectors form a subspace and often transfer across tasks. Our results demonstrate that LMs learn interpretable strategies for representing symbolic knowledge in context, and that studying context activations is a fruitful direction for understanding LM cognition.

Cite

Text

Feng and Steinhardt. "How Do Language Models Bind Entities in Context?." International Conference on Learning Representations, 2024.

Markdown

[Feng and Steinhardt. "How Do Language Models Bind Entities in Context?." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/feng2024iclr-language/)

BibTeX

@inproceedings{feng2024iclr-language,
  title     = {{How Do Language Models Bind Entities in Context?}},
  author    = {Feng, Jiahai and Steinhardt, Jacob},
  booktitle = {International Conference on Learning Representations},
  year      = {2024},
  url       = {https://mlanthology.org/iclr/2024/feng2024iclr-language/}
}