A Closer Look at In-Context Learning Under Distribution Shifts
Abstract
In-context learning, a capability that enables a model to learn from input examples on the fly without necessitating weight updates, is a defining characteristic of large language models. In this work, we follow the setting proposed in Garg et al. to better understand the generality and limitations of in-context learning from the lens of the simple yet fundamental task of linear regression. The key question we aim to address is: Are transformers more adept than some natural and simpler architectures at performing in-context learning under varying distribution shifts? To compare transformers, we propose to use a simple architecture based on set-based Multi-Layer Perceptrons (MLPs). We find that both transformers and set-based MLPs exhibit in-context learning under in-distribution evaluations, but transformers more closely emulate the performance of ordinary least squares (OLS). Transformers also display better resilience to mild distribution shifts, where set-based MLPs falter. However, under severe distribution shifts, both models' in-context learning abilities diminish.
Cite
Text
Ahuja and Lopez-Paz. "A Closer Look at In-Context Learning Under Distribution Shifts." ICML 2023 Workshops: ES-FoMO, 2023.Markdown
[Ahuja and Lopez-Paz. "A Closer Look at In-Context Learning Under Distribution Shifts." ICML 2023 Workshops: ES-FoMO, 2023.](https://mlanthology.org/icmlw/2023/ahuja2023icmlw-closer/)BibTeX
@inproceedings{ahuja2023icmlw-closer,
title = {{A Closer Look at In-Context Learning Under Distribution Shifts}},
author = {Ahuja, Kartik and Lopez-Paz, David},
booktitle = {ICML 2023 Workshops: ES-FoMO},
year = {2023},
url = {https://mlanthology.org/icmlw/2023/ahuja2023icmlw-closer/}
}