Detrimental Memories in Transfer Learning

Abstract

The source domain in transfer learning provides essential features that enable effective and data-efficient learning on the target task. Typically, the finetuning process does not explicitly account for how the knowledge about the source domain interacts with the target task. We demonstrate how that knowledge can interfere with the target task leading to negative transfer. Specifically, certain memories about the source domain can distract the finetuned model in certain inputs. We provide a method to analyze those memories in typical foundational models and to surface potential failure cases of those models. This analysis helps model developers explore remedies for those failure cases. Our results can be reproduced at https://github.com/AmAlnouri-JKU/TL_Interference

Cite

Text

Alnouri et al. "Detrimental Memories in Transfer Learning." ICML 2024 Workshops: TF2M, 2024.

Markdown

[Alnouri et al. "Detrimental Memories in Transfer Learning." ICML 2024 Workshops: TF2M, 2024.](https://mlanthology.org/icmlw/2024/alnouri2024icmlw-detrimental/)

BibTeX

@inproceedings{alnouri2024icmlw-detrimental,
  title     = {{Detrimental Memories in Transfer Learning}},
  author    = {Alnouri, Amal and Wroge, Timothy J and Alsallakh, Bilal},
  booktitle = {ICML 2024 Workshops: TF2M},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/alnouri2024icmlw-detrimental/}
}