Project and Probe: Sample-Efficient Adaptation by Interpolating Orthogonal Features
Abstract
Transfer learning with a small amount of target data is an effective and common approach to adapting a pre-trained model to distribution shifts. In some situations, target data labels may be expensive to obtain, so we may only have access to a limited number of target data points. To make the most of a very small target dataset, we propose a lightweight, sample-efficient approach that learns a diverse set of features and adapts to a target distribution by interpolating these features. Our approach, Project and Probe (Pro$^2$), first learns a linear projection that maps a pre-trained embedding onto orthogonal directions while being predictive of labels in the source dataset. The goal of this step is to learn a variety of predictive features, so that at least some of them remain useful after distribution shift. Pro$^2$ then learns a linear classifier on top of these projected features using a small target dataset. Theoretically, we find that Pro$^2$ results in more sample-efficient generalization by inducing a favorable bias-variance tradeoff. Our experiments on four datasets, with multiple distribution shift settings for each, show that Pro$^2$ improves performance by 5-15% when given limited target data compared to prior methods such as standard linear probing.
Cite
Text
Chen et al. "Project and Probe: Sample-Efficient Adaptation by Interpolating Orthogonal Features." International Conference on Learning Representations, 2024.Markdown
[Chen et al. "Project and Probe: Sample-Efficient Adaptation by Interpolating Orthogonal Features." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/chen2024iclr-project/)BibTeX
@inproceedings{chen2024iclr-project,
title = {{Project and Probe: Sample-Efficient Adaptation by Interpolating Orthogonal Features}},
author = {Chen, Annie S and Lee, Yoonho and Setlur, Amrith and Levine, Sergey and Finn, Chelsea},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/chen2024iclr-project/}
}