Foundation Models for Semantic Novelty in Reinforcement Learning

Abstract

Effectively exploring the environment is a key challenge in reinforcement learning (RL). We address this challenge by defining a novel intrinsic reward based on a foundation model, such as contrastive language image pretraining (CLIP), which can encode a wealth of domain-independent semantic visual-language knowledge about the world. Specifically, our intrinsic reward is defined based on pre-trained CLIP embeddings without any fine-tuning or learning on the target RL task. We demonstrate that CLIP-based intrinsic rewards can drive exploration towards semantically meaningful states and outperform state-of-the-art methods in challenging sparse-reward procedurally-generated environments.

Cite

Text

Gupta et al. "Foundation Models for Semantic Novelty in Reinforcement Learning." NeurIPS 2022 Workshops: FMDM, 2022.

Markdown

[Gupta et al. "Foundation Models for Semantic Novelty in Reinforcement Learning." NeurIPS 2022 Workshops: FMDM, 2022.](https://mlanthology.org/neuripsw/2022/gupta2022neuripsw-foundation/)

BibTeX

@inproceedings{gupta2022neuripsw-foundation,
  title     = {{Foundation Models for Semantic Novelty in Reinforcement Learning}},
  author    = {Gupta, Tarun and Karkus, Peter and Che, Tong and Xu, Danfei and Pavone, Marco},
  booktitle = {NeurIPS 2022 Workshops: FMDM},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/gupta2022neuripsw-foundation/}
}