Considerations for Distribution Shift Robustness in Health
Abstract
When analyzing robustness of predictive models under distribution shift, many works focus on tackling generalization in the presence of spurious correlations. In this case, one typically makes use of covariates or environment indicators to enforce independencies in learned models to guarantee generalization under various distribution shifts. In this work, we analyze a class of distribution shifts, where such independencies are not desirable, as there is a causal association between covariates and outcomes of interest. This case is common in the health space where covariates can be causally, as opposed to spuriously, related to outcomes of interest. We formalize this setting and relate it to common distribution shift settings from the literature. We theoretically show why standard supervised learning and invariant learning will not yield robust predictors in this case, while including the causal covariates into the prediction model can recover robustness. We demonstrate our theoretical findings in experiments on both synthetic and real data.
Cite
Text
Blaas et al. "Considerations for Distribution Shift Robustness in Health." ICLR 2023 Workshops: TML4H, 2023.Markdown
[Blaas et al. "Considerations for Distribution Shift Robustness in Health." ICLR 2023 Workshops: TML4H, 2023.](https://mlanthology.org/iclrw/2023/blaas2023iclrw-considerations/)BibTeX
@inproceedings{blaas2023iclrw-considerations,
title = {{Considerations for Distribution Shift Robustness in Health}},
author = {Blaas, Arno and Miller, Andrew and Zappella, Luca and Jacobsen, Joern-Henrik and Heinze-Deml, Christina},
booktitle = {ICLR 2023 Workshops: TML4H},
year = {2023},
url = {https://mlanthology.org/iclrw/2023/blaas2023iclrw-considerations/}
}