Just Mix Once: Mixing Samples with Implicit Group Distribution
Abstract
Recent work has unveiled how average generalization frequently relies on superficial patterns in data. The consequences are brittle models with poor performance in the presence of domain shift in group distribution at test time. When the subgroups in the training data are known, we can use tools from robust optimization to tackle the problem. However, group annotation and identification are time-consuming tasks, especially on large datasets. A recent line of research~\cite{liu2021just} is trying to solve this problem with implicit group distribution at training time, leveraging self-supervision and oversampling to improve generalization on minority groups. Following such ideas, we propose a new class-conditional variant of MixUp~\cite{zhang2017mixup} for worst-group generalization, augmenting the training distribution with a continuous distribution of groups. Our method, called Just Mix Once (JM1), is domain-agnostic, computationally efficient, and performs on par or better than the state-of-the-art on worst-group generalization.
Cite
Text
Giannone et al. "Just Mix Once: Mixing Samples with Implicit Group Distribution." NeurIPS 2021 Workshops: DistShift, 2021.Markdown
[Giannone et al. "Just Mix Once: Mixing Samples with Implicit Group Distribution." NeurIPS 2021 Workshops: DistShift, 2021.](https://mlanthology.org/neuripsw/2021/giannone2021neuripsw-just/)BibTeX
@inproceedings{giannone2021neuripsw-just,
title = {{Just Mix Once: Mixing Samples with Implicit Group Distribution}},
author = {Giannone, Giorgio and Havrylov, Serhii and Massiah, Jordan and Yilmaz, Emine and Jiao, Yunlong},
booktitle = {NeurIPS 2021 Workshops: DistShift},
year = {2021},
url = {https://mlanthology.org/neuripsw/2021/giannone2021neuripsw-just/}
}