CDS: Cross-Domain Self-Supervised Pre-Training
Abstract
We present a two-stage pre-training approach that improves the generalization ability of standard single-domain pre-training. While standard pre-training on a single large dataset (such as ImageNet) can provide a good initial representation for transfer learning tasks, this approach may result in biased representations that impact the success of learning with new multi-domain data (e.g., different artistic styles) via methods like domain adaptation. We propose a novel pre-training approach called Cross-Domain Self-supervision (CDS), which directly employs unlabeled multi-domain data for downstream domain transfer tasks. Our approach uses self-supervision not only within a single domain but also across domains. In-domain instance discrimination is used to learn discriminative features on new data in a domain-adaptive manner, while cross-domain matching is used to learn domain-invariant features. We apply our method as a second pre-training step (after ImageNet pre-training), resulting in a significant target accuracy boost to diverse domain transfer tasks compared to standard one-stage pre-training.
Cite
Text
Kim et al. "CDS: Cross-Domain Self-Supervised Pre-Training." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.00899Markdown
[Kim et al. "CDS: Cross-Domain Self-Supervised Pre-Training." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/kim2021iccv-cds/) doi:10.1109/ICCV48922.2021.00899BibTeX
@inproceedings{kim2021iccv-cds,
title = {{CDS: Cross-Domain Self-Supervised Pre-Training}},
author = {Kim, Donghyun and Saito, Kuniaki and Oh, Tae-Hyun and Plummer, Bryan A. and Sclaroff, Stan and Saenko, Kate},
booktitle = {International Conference on Computer Vision},
year = {2021},
pages = {9123-9132},
doi = {10.1109/ICCV48922.2021.00899},
url = {https://mlanthology.org/iccv/2021/kim2021iccv-cds/}
}