NICO++: Towards Better Benchmarking for Domain Generalization

Abstract

Despite the remarkable performance that modern deep neural networks have achieved on independent and identically distributed (I.I.D.) data, they can crash under distribution shifts. Most current evaluation methods for domain generalization (DG) adopt the leave-one-out strategy as a compromise on the limited number of domains. We propose a large-scale benchmark with extensive labeled domains named NICO++ along with more rational evaluation methods for comprehensively evaluating DG algorithms. To evaluate DG datasets, we propose two metrics to quantify covariate shift and concept shift, respectively. Two novel generalization bounds from the perspective of data construction are proposed to prove that limited concept shift and significant covariate shift favor the evaluation capability for generalization. Through extensive experiments, NICO++ shows its superior evaluation capability compared with current DG datasets and its contribution in alleviating unfairness caused by the leak of oracle knowledge in model selection.

Cite

Text

Zhang et al. "NICO++: Towards Better Benchmarking for Domain Generalization." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.01539

Markdown

[Zhang et al. "NICO++: Towards Better Benchmarking for Domain Generalization." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/zhang2023cvpr-nico/) doi:10.1109/CVPR52729.2023.01539

BibTeX

@inproceedings{zhang2023cvpr-nico,
  title     = {{NICO++: Towards Better Benchmarking for Domain Generalization}},
  author    = {Zhang, Xingxuan and He, Yue and Xu, Renzhe and Yu, Han and Shen, Zheyan and Cui, Peng},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {16036-16047},
  doi       = {10.1109/CVPR52729.2023.01539},
  url       = {https://mlanthology.org/cvpr/2023/zhang2023cvpr-nico/}
}