Towards General-Purpose In-Context Learning Agents

Abstract

Reinforcement Learning (RL) algorithms are usually hand-crafted, driven by the research and engineering of humans. An alternative approach is to automate this research process via meta-learning. A particularly ambitious objective is to automatically discover new RL algorithms from scratch that use in-context learning to learn-how-to-learn entirely from data while also generalizing to a wide range of environments. Those RL algorithms are implemented entirely in neural networks, by conditioning on previous experience from the environment, without any explicit optimization-based routine at meta-test time. To achieve generalization, this requires a broad task distribution of diverse and challenging environments. Our Transformer-based Generally Learning Agents (GLAs) are an important first step in this direction. Our GLAs are meta-trained using supervised learning techniques on an offline dataset with experiences from RL environments that is augmented with random projections to generate task diversity. During meta-testing our agents perform in-context meta-RL on entirely different robotic control problems such as Reacher, Cartpole, or HalfCheetah that were not in the meta-training distribution.

Cite

Text

Kirsch et al. "Towards General-Purpose In-Context Learning Agents." NeurIPS 2023 Workshops: R0-FoMo, 2023.

Markdown

[Kirsch et al. "Towards General-Purpose In-Context Learning Agents." NeurIPS 2023 Workshops: R0-FoMo, 2023.](https://mlanthology.org/neuripsw/2023/kirsch2023neuripsw-generalpurpose-c/)

BibTeX

@inproceedings{kirsch2023neuripsw-generalpurpose-c,
  title     = {{Towards General-Purpose In-Context Learning Agents}},
  author    = {Kirsch, Louis and Harrison, James and Freeman, C. and Sohl-Dickstein, Jascha and Schmidhuber, Jürgen},
  booktitle = {NeurIPS 2023 Workshops: R0-FoMo},
  year      = {2023},
  url       = {https://mlanthology.org/neuripsw/2023/kirsch2023neuripsw-generalpurpose-c/}
}