Limits of Model Selection Under Transfer Learning

Abstract

Theoretical studies on \emph{transfer learning} (or \emph{domain adaptation}) have so far focused on situations with a known hypothesis class or \emph{model}; however in practice, some amount of model selection is usually involved, often appearing under the umbrella term or \emph{hyperparameter-tuning}: for example, one may think of the problem of \emph{tuning} for the right neural network architecture towards a target task, while leveraging data from a related \emph{source} task. Now, in addition to the usual tradeoffs on approximation vs estimation errors involved in model selection, this problem brings in a new complexity term, namely, the \emph{transfer distance} between source and target distributions, which is known to vary with the choice of hypothesis class. We present a first study of this problem, focusing on classification; in particular, the analysis reveals some remarkable phenomena: \emph{adaptive rates}, i.e., those achievable with no distributional information, can be arbitrarily slower than \emph{oracle rates}, i.e., when given knowledge on \emph{distances}

Cite

Text

Hanneke et al. "Limits of Model Selection Under Transfer Learning." Conference on Learning Theory, 2023.

Markdown

[Hanneke et al. "Limits of Model Selection Under Transfer Learning." Conference on Learning Theory, 2023.](https://mlanthology.org/colt/2023/hanneke2023colt-limits/)

BibTeX

@inproceedings{hanneke2023colt-limits,
  title     = {{Limits of Model Selection Under Transfer Learning}},
  author    = {Hanneke, Steve and Kpotufe, Samory and Mahdaviyeh, Yasaman},
  booktitle = {Conference on Learning Theory},
  year      = {2023},
  pages     = {5781-5812},
  volume    = {195},
  url       = {https://mlanthology.org/colt/2023/hanneke2023colt-limits/}
}