Increasing Brain-LLM Alignment via Information-Theoretic Compression

Abstract

Recent work has discovered similarities between learned representations in large language models (LLMs) and human brain activity during language processing. However, it remains unclear what information LLM and brain representations share. In this work, inspired by a notion that brain data may include information not captured by LLMs, we apply an information bottleneck method to generate compressed representations of fMRI data. For certain brain regions in the frontal cortex, we find that compressing brain representations by a small amount increases their similarity to both BERT and GPT2 embeddings. Thus, our method not only improves LLM-brain alignment scores but also suggests important characteristics about the amount of information captured by each representation scheme.

Cite

Text

Tucker and Tuckute. "Increasing Brain-LLM Alignment via Information-Theoretic Compression." NeurIPS 2023 Workshops: UniReps, 2023.

Markdown

[Tucker and Tuckute. "Increasing Brain-LLM Alignment via Information-Theoretic Compression." NeurIPS 2023 Workshops: UniReps, 2023.](https://mlanthology.org/neuripsw/2023/tucker2023neuripsw-increasing/)

BibTeX

@inproceedings{tucker2023neuripsw-increasing,
  title     = {{Increasing Brain-LLM Alignment via Information-Theoretic Compression}},
  author    = {Tucker, Mycal and Tuckute, Greta},
  booktitle = {NeurIPS 2023 Workshops: UniReps},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/tucker2023neuripsw-increasing/}
}