Representation Learning as Finding Necessary and Sufficient Causes
Abstract
Representation learning constructs low-dimensional representations to summarize essential features of high-dimensional data. This learning problem is often approached by describing various desiderata associated with learned representations; e.g., that they be non-spurious or efficient. It can be challenging, however, to turn these intuitive desiderata into formal criteria that can be measured and enhanced based on observed data. In this paper, we take a causal perspective on representation learning, formalizing non-spuriousness and efficiency (in supervised representation learning) using counterfactual quantities and observable consequences of causal assertions. This yields computable metrics that can be used to assess the degree to which representations satisfy the desiderata of interest and learn non-spurious representations from single observational datasets.
Cite
Text
Wang and Jordan. "Representation Learning as Finding Necessary and Sufficient Causes." ICML 2022 Workshops: SCIS, 2022.Markdown
[Wang and Jordan. "Representation Learning as Finding Necessary and Sufficient Causes." ICML 2022 Workshops: SCIS, 2022.](https://mlanthology.org/icmlw/2022/wang2022icmlw-representation/)BibTeX
@inproceedings{wang2022icmlw-representation,
title = {{Representation Learning as Finding Necessary and Sufficient Causes}},
author = {Wang, Yixin and Jordan, Michael},
booktitle = {ICML 2022 Workshops: SCIS},
year = {2022},
url = {https://mlanthology.org/icmlw/2022/wang2022icmlw-representation/}
}