CODA: A Correlation-Oriented Disentanglement and Augmentation Modeling Scheme for Better Resisting Subpopulation Shifts
Abstract
Data-driven models learned often struggle to generalize due to widespread subpopulation shifts, especially the presence of both spurious correlations and group imbalance (SC-GI). To learn models more powerful for defending against SC-GI, we propose a {\bf Correlation-Oriented Disentanglement and Augmentation (CODA)} modeling scheme, which includes two unique developments: (1) correlation-oriented disentanglement and (2) strategic sample augmentation with reweighted consistency (RWC) loss. In (1), a bi-branch encoding process is developed to enable the disentangling of variant and invariant correlations by coordinating with a decoy classifier and the decoder reconstruction. In (2), a strategic sample augmentation based on disentangled latent features with RWC loss is designed to reinforce the training of a more generalizable model. The effectiveness of CODA is verified by benchmarking against a set of SOTA models in terms of worst-group accuracy and maximum group accuracy gap based on two famous datasets, ColoredMNIST and CelebA.
Cite
Text
Ou and Zhang. "CODA: A Correlation-Oriented Disentanglement and Augmentation Modeling Scheme for Better Resisting Subpopulation Shifts." Neural Information Processing Systems, 2024. doi:10.52202/079017-3258Markdown
[Ou and Zhang. "CODA: A Correlation-Oriented Disentanglement and Augmentation Modeling Scheme for Better Resisting Subpopulation Shifts." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/ou2024neurips-coda/) doi:10.52202/079017-3258BibTeX
@inproceedings{ou2024neurips-coda,
title = {{CODA: A Correlation-Oriented Disentanglement and Augmentation Modeling Scheme for Better Resisting Subpopulation Shifts}},
author = {Ou, Ziquan and Zhang, Zijun},
booktitle = {Neural Information Processing Systems},
year = {2024},
doi = {10.52202/079017-3258},
url = {https://mlanthology.org/neurips/2024/ou2024neurips-coda/}
}