$i$-Mix: A Domain-Agnostic Strategy for Contrastive Representation Learning

Abstract

Contrastive representation learning has shown to be effective to learn representations from unlabeled data. However, much progress has been made in vision domains relying on data augmentations carefully designed using domain knowledge. In this work, we propose i-Mix, a simple yet effective domain-agnostic regularization strategy for improving contrastive representation learning. We cast contrastive learning as training a non-parametric classifier by assigning a unique virtual class to each data in a batch. Then, data instances are mixed in both the input and virtual label spaces, providing more augmented data during training. In experiments, we demonstrate that i-Mix consistently improves the quality of learned representations across domains, including image, speech, and tabular data. Furthermore, we confirm its regularization effect via extensive ablation studies across model and dataset sizes. The code is available at https://github.com/kibok90/imix.

Cite

Text

Lee et al. "$i$-Mix: A Domain-Agnostic Strategy for Contrastive Representation Learning." International Conference on Learning Representations, 2021.

Markdown

[Lee et al. "$i$-Mix: A Domain-Agnostic Strategy for Contrastive Representation Learning." International Conference on Learning Representations, 2021.](https://mlanthology.org/iclr/2021/lee2021iclr-imix/)

BibTeX

@inproceedings{lee2021iclr-imix,
  title     = {{$i$-Mix: A Domain-Agnostic Strategy for Contrastive Representation Learning}},
  author    = {Lee, Kibok and Zhu, Yian and Sohn, Kihyuk and Li, Chun-Liang and Shin, Jinwoo and Lee, Honglak},
  booktitle = {International Conference on Learning Representations},
  year      = {2021},
  url       = {https://mlanthology.org/iclr/2021/lee2021iclr-imix/}
}