Multimodal Neural Networks Better Explain Multivoxel Patterns in the Hippocampus

Abstract

The human hippocampus possesses "concept cells", neurons that fire when presented with stimuli belonging to a specific concept, regardless of the modality. Recently, similar concept cells were discovered in a multimodal network called CLIP [1].Here, we ask whether CLIP can explain the fMRI activity of the human hippocampus better than a purely visual (or linguistic) model. We extend our analysis to a range of publicly available uni- and multi-modal models. We demonstrate that ``multimodality'' stands out as a key component when assessing the ability of a network to explain the multivoxel activity in the hippocampus.

Cite

Text

Choksi et al. "Multimodal Neural Networks Better Explain Multivoxel Patterns in the Hippocampus." NeurIPS 2021 Workshops: SVRHM, 2021.

Markdown

[Choksi et al. "Multimodal Neural Networks Better Explain Multivoxel Patterns in the Hippocampus." NeurIPS 2021 Workshops: SVRHM, 2021.](https://mlanthology.org/neuripsw/2021/choksi2021neuripsw-multimodal/)

BibTeX

@inproceedings{choksi2021neuripsw-multimodal,
  title     = {{Multimodal Neural Networks Better Explain Multivoxel Patterns in the Hippocampus}},
  author    = {Choksi, Bhavin and Mozafari, Milad and VanRullen, Rufin and Reddy, Leila},
  booktitle = {NeurIPS 2021 Workshops: SVRHM},
  year      = {2021},
  url       = {https://mlanthology.org/neuripsw/2021/choksi2021neuripsw-multimodal/}
}