On Counterfactual Inference with Unobserved Confounding

Abstract

Given an observational study with $n$ independent but heterogeneous units and one $p$-dimensional sample per unit containing covariates, interventions, and outcomes, our goal is to learn counterfactual distribution for each unit. We consider studies with unobserved confounding which introduces statistical biases between interventions and outcomes as well as exacerbates the heterogeneity across units. Modeling the underlying joint distribution as an exponential family and under suitable conditions, we reduce learning the $n$ unit-level counterfactual distributions to learning $n$ exponential family distributions with heterogeneous parameters and only one sample per distribution. We introduce a convex objective that pools all $n$ samples to jointly learn all $n$ parameters and provide a unit-wise mean squared error bound that scales linearly with the metric entropy of the parameter space. For example, when the parameters are $s$-sparse linear combination of $k$ known vectors, the error is $O(s\log k/p)$. En route, we derive sufficient conditions for compactly supported distributions to satisfy the logarithmic Sobolev inequality.

Cite

Text

Shah et al. "On Counterfactual Inference with Unobserved Confounding." NeurIPS 2022 Workshops: CML4Impact, 2022.

Markdown

[Shah et al. "On Counterfactual Inference with Unobserved Confounding." NeurIPS 2022 Workshops: CML4Impact, 2022.](https://mlanthology.org/neuripsw/2022/shah2022neuripsw-counterfactual/)

BibTeX

@inproceedings{shah2022neuripsw-counterfactual,
  title     = {{On Counterfactual Inference with Unobserved Confounding}},
  author    = {Shah, Abhin and Dwivedi, Raaz and Shah, Devavrat and Wornell, Gregory},
  booktitle = {NeurIPS 2022 Workshops: CML4Impact},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/shah2022neuripsw-counterfactual/}
}