Tailored Overlap for Learning Under Distribution Shift
Abstract
Distributional overlap is a critical determinant of learnability in domain adaptation. The standard theory quantifies overlap in terms of $\chi^2$ divergence, as this factors directly into variance and generalization bounds agnostic to the functional form of the $Y$-$X$ relationship. However, in many modern settings, we cannot afford this agnosticism; we often wish to transfer across distributions with disjoint support, where these standard divergence measures are infinite. In this note, we argue that ``tailored'' divergences that are restricted to measuring overlap in a particular function class are more appropriate. We show how $\chi^2$ (and other) divergences can be generalized to this restricted function class setting via a variational representation, and use this to motivate balancing weight-based methods that have been proposed before, but, we believe, should be more widely used.
Cite
Text
Bruns-Smith et al. "Tailored Overlap for Learning Under Distribution Shift." NeurIPS 2022 Workshops: DistShift, 2022.Markdown
[Bruns-Smith et al. "Tailored Overlap for Learning Under Distribution Shift." NeurIPS 2022 Workshops: DistShift, 2022.](https://mlanthology.org/neuripsw/2022/brunssmith2022neuripsw-tailored/)BibTeX
@inproceedings{brunssmith2022neuripsw-tailored,
title = {{Tailored Overlap for Learning Under Distribution Shift}},
author = {Bruns-Smith, David and D'Amour, Alexander and Feller, Avi and Yadlowsky, Steve},
booktitle = {NeurIPS 2022 Workshops: DistShift},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/brunssmith2022neuripsw-tailored/}
}