Representation with a Capital 'r': Measuring Functional Alignment with Causal Perturbation
Abstract
Ambiguity surrounding the term 'representation' in biological and artificial neural systems hampers our ability to assess their alignment. In this paper, we draw a critical distinction between two notions of representation: the conventional 'representation' as a mere encoding, and 'Representation-with-a capital-R', which entails functional use within a system. We argue that while current methods in neuroscience and artificial intelligence often focus on the former, advancing our understanding requires a shift toward the latter. We critique existing linking methods such as representational similarity analysis and encoding models, highlighting key limitations in their ability to capture functional correspondence. We then propose an updated paradigm that involves specifying explicit models of information readout and testing their functional properties using causal perturbation. This framework, which treats neural networks as a new species of model organism, may help reveal the principles governing functional representations in both biological and artificial systems.
Cite
Text
Prince et al. "Representation with a Capital 'r': Measuring Functional Alignment with Causal Perturbation." NeurIPS 2024 Workshops: UniReps, 2024.Markdown
[Prince et al. "Representation with a Capital 'r': Measuring Functional Alignment with Causal Perturbation." NeurIPS 2024 Workshops: UniReps, 2024.](https://mlanthology.org/neuripsw/2024/prince2024neuripsw-representation/)BibTeX
@inproceedings{prince2024neuripsw-representation,
title = {{Representation with a Capital 'r': Measuring Functional Alignment with Causal Perturbation}},
author = {Prince, Jacob S. and Alvarez, George A. and Konkle, Talia},
booktitle = {NeurIPS 2024 Workshops: UniReps},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/prince2024neuripsw-representation/}
}