Context-Aware Self-Adaptation for Domain Generalization
Abstract
Domain generalization aims at developing suitable learning algorithms in source training domains such that the model learned can generalize well on a different unseen testing domain. We present a novel two-stage approach called Context-Aware Self-Adaptation (CASA) for domain generalization. CASA simulates an approximate meta-generalization scenario and incorporates a self-adaptation module to adjust pre-trained meta-source models to the meta-target domains while maintaining their predictive capability on the meta-source domains. The core concept of self-adaptation involves leveraging contextual information, such as the mean of mini-batch features, as domain knowledge to automatically adapt a model trained in the first stage to new contexts in the second stage. Lastly, we utilize an ensemble of multiple meta-source models to perform inference on the testing domain. Experimental results demonstrate that our proposed method achieves state-of-the-art performance on standard benchmarks.
Cite
Text
Yan and Guo. "Context-Aware Self-Adaptation for Domain Generalization." ICML 2023 Workshops: AdvML-Frontiers, 2023.Markdown
[Yan and Guo. "Context-Aware Self-Adaptation for Domain Generalization." ICML 2023 Workshops: AdvML-Frontiers, 2023.](https://mlanthology.org/icmlw/2023/yan2023icmlw-contextaware/)BibTeX
@inproceedings{yan2023icmlw-contextaware,
title = {{Context-Aware Self-Adaptation for Domain Generalization}},
author = {Yan, Hao and Guo, Yuhong},
booktitle = {ICML 2023 Workshops: AdvML-Frontiers},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/yan2023icmlw-contextaware/}
}