Case-Based Reasoning for Better Generalization in Textual Reinforcement Learning

Abstract

Text-based games (TBG) have emerged as promising environments for driving research in grounded language understanding and studying problems like generalization and sample efficiency. Several deep reinforcement learning (RL) methods with varying architectures and learning schemes have been proposed for TBGs. However, these methods fail to generalize efficiently, especially under distributional shifts. In a departure from deep RL approaches, in this paper, we propose a general method inspired by case-based reasoning to train agents and generalize out of the training distribution. The case-based reasoner collects instances of positive experiences from the agent's interaction with the world and later reuses the collected experiences to act efficiently. The method can be used in conjunction with any existing on-policy neural agent introduced in the literature for TBGs. Our experiments show that the proposed approach consistently improves existing methods, obtains good out-of-distribution generalization and achieves new state-of-the-art results on widely used environments.

Cite

Text

Atzeni et al. "Case-Based Reasoning for Better Generalization in Textual Reinforcement Learning." International Conference on Learning Representations, 2022.

Markdown

[Atzeni et al. "Case-Based Reasoning for Better Generalization in Textual Reinforcement Learning." International Conference on Learning Representations, 2022.](https://mlanthology.org/iclr/2022/atzeni2022iclr-casebased/)

BibTeX

@inproceedings{atzeni2022iclr-casebased,
  title     = {{Case-Based Reasoning for Better Generalization in Textual Reinforcement Learning}},
  author    = {Atzeni, Mattia and Dhuliawala, Shehzaad Zuzar and Murugesan, Keerthiram and Sachan, Mrinmaya},
  booktitle = {International Conference on Learning Representations},
  year      = {2022},
  url       = {https://mlanthology.org/iclr/2022/atzeni2022iclr-casebased/}
}