Kernel Instrumental Variable Regression

Abstract

Instrumental variable (IV) regression is a strategy for learning causal relationships in observational data. If measurements of input X and output Y are confounded, the causal relationship can nonetheless be identified if an instrumental variable Z is available that influences X directly, but is conditionally independent of Y given X and the unmeasured confounder. The classic two-stage least squares algorithm (2SLS) simplifies the estimation problem by modeling all relationships as linear functions. We propose kernel instrumental variable regression (KIV), a nonparametric generalization of 2SLS, modeling relations among X, Y, and Z as nonlinear functions in reproducing kernel Hilbert spaces (RKHSs). We prove the consistency of KIV under mild assumptions, and derive conditions under which convergence occurs at the minimax optimal rate for unconfounded, single-stage RKHS regression. In doing so, we obtain an efficient ratio between training sample sizes used in the algorithm's first and second stages. In experiments, KIV outperforms state of the art alternatives for nonparametric IV regression.

Cite

Text

Singh et al. "Kernel Instrumental Variable Regression." Neural Information Processing Systems, 2019.

Markdown

[Singh et al. "Kernel Instrumental Variable Regression." Neural Information Processing Systems, 2019.](https://mlanthology.org/neurips/2019/singh2019neurips-kernel/)

BibTeX

@inproceedings{singh2019neurips-kernel,
  title     = {{Kernel Instrumental Variable Regression}},
  author    = {Singh, Rahul and Sahani, Maneesh and Gretton, Arthur},
  booktitle = {Neural Information Processing Systems},
  year      = {2019},
  pages     = {4593-4605},
  url       = {https://mlanthology.org/neurips/2019/singh2019neurips-kernel/}
}