From Causal to Concept-Based Representation Learning
Abstract
To build intelligent machine learning systems, there are two broad approaches. One approach is to build inherently interpretable models, as endeavored by the growing field of causal representation learning. The other approach is to build highly-performant foundation models and then invest efforts into understanding how they work. In this work, we relate these two approaches and study how to learn human-interpretable concepts from data. Weaving together ideas from both fields, we formally define a notion of concepts and prove that they can be identifiably recovered from diverse data. Experiments on synthetic data, CLIP models and large language models show the utility of our unified approach.
Cite
Text
Rajendran et al. "From Causal to Concept-Based Representation Learning." NeurIPS 2024 Workshops: CALM, 2024.Markdown
[Rajendran et al. "From Causal to Concept-Based Representation Learning." NeurIPS 2024 Workshops: CALM, 2024.](https://mlanthology.org/neuripsw/2024/rajendran2024neuripsw-causal/)BibTeX
@inproceedings{rajendran2024neuripsw-causal,
title = {{From Causal to Concept-Based Representation Learning}},
author = {Rajendran, Goutham and Buchholz, Simon and Aragam, Bryon and Schölkopf, Bernhard and Ravikumar, Pradeep Kumar},
booktitle = {NeurIPS 2024 Workshops: CALM},
year = {2024},
url = {https://mlanthology.org/neuripsw/2024/rajendran2024neuripsw-causal/}
}