Causal Balancing for Domain Generalization
Abstract
While machine learning models rapidly advance the state-of-the-art on various real-world tasks, out-of-domain (OOD) generalization remains a challenging problem given the vulnerability of these models to spurious correlations. We propose a causally-motivated balanced mini-batch sampling strategy to train robust classifiers that is minimax optimal across a diverse enough environment space, by utilizing multiple training sets from different environments. We provide an identifiability guarantee of the latent covariates in the proposed causal graph and show that our proposed approach samples train data from a balanced, spurious-free distribution under an ideal scenario. Experiments are conducted on three domain generalization datasets, demonstrating empirically that our balanced mini-batch sampling strategy improves the performance of four different established domain generalization model baselines compared to the random mini-batch sampling strategy.
Cite
Text
Wang et al. "Causal Balancing for Domain Generalization." ICML 2022 Workshops: SCIS, 2022.Markdown
[Wang et al. "Causal Balancing for Domain Generalization." ICML 2022 Workshops: SCIS, 2022.](https://mlanthology.org/icmlw/2022/wang2022icmlw-causal/)BibTeX
@inproceedings{wang2022icmlw-causal,
title = {{Causal Balancing for Domain Generalization}},
author = {Wang, Xinyi and Saxon, Michael and Li, Jiachen and Zhang, Hongyang and Zhang, Kun and Wang, William Yang},
booktitle = {ICML 2022 Workshops: SCIS},
year = {2022},
url = {https://mlanthology.org/icmlw/2022/wang2022icmlw-causal/}
}