Meta- (out-of-Context) Learning in Neural Networks

Abstract

Brown et al. (2020) famously introduced the phenomenon of in-context learning in large language models (LLMs). We establish the existence of a phenomenon we call **meta-out-of-context learning (meta-OCL)** via carefully designed synthetic experiments with LLMs. Our results suggest that meta-OCL leads LLMs to more readily “internalize” the semantic content of text that is, *or appears to be*, broadly useful (such as true statements, or text from authoritative sources) and use it in appropriate circumstances. We further demonstrate meta-OCL in a synthetic computer vision setting, and propose two hypotheses for the emergence of meta-OCL: one relying on the way models store knowledge in their parameters, and another suggesting that the implicit gradient alignment bias of gradient-descent-based optimizers may be responsible. Finally, we reflect on what our results might imply about capabilities of future AI systems, and discuss potential risks. Our code is available at https://github.com/krasheninnikov/internalization.

Cite

Text

Krasheninnikov et al. "Meta- (out-of-Context) Learning in Neural Networks." NeurIPS 2023 Workshops: R0-FoMo, 2023.

Markdown

[Krasheninnikov et al. "Meta- (out-of-Context) Learning in Neural Networks." NeurIPS 2023 Workshops: R0-FoMo, 2023.](https://mlanthology.org/neuripsw/2023/krasheninnikov2023neuripsw-meta/)

BibTeX

@inproceedings{krasheninnikov2023neuripsw-meta,
  title     = {{Meta- (out-of-Context) Learning in Neural Networks}},
  author    = {Krasheninnikov, Dmitrii and Krasheninnikov, Egor and Mlodozeniec, Bruno and Krueger, David},
  booktitle = {NeurIPS 2023 Workshops: R0-FoMo},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/krasheninnikov2023neuripsw-meta/}
}