Generalization to New Actions in Reinforcement Learning
Abstract
A fundamental trait of intelligence is the ability to achieve goals in the face of novel circumstances, such as making decisions from new action choices. However, standard reinforcement learning assumes a fixed set of actions and requires expensive retraining when given a new action set. To make learning agents more adaptable, we introduce the problem of zero-shot generalization to new actions. We propose a two-stage framework where the agent first infers action representations from action information acquired separately from the task. A policy flexible to varying action sets is then trained with generalization objectives. We benchmark generalization on sequential tasks, such as selecting from an unseen tool-set to solve physical reasoning puzzles and stacking towers with novel 3D shapes. Videos and code are available at https://sites.google.com/view/action-generalization.
Cite
Text
Jain et al. "Generalization to New Actions in Reinforcement Learning." International Conference on Machine Learning, 2020.Markdown
[Jain et al. "Generalization to New Actions in Reinforcement Learning." International Conference on Machine Learning, 2020.](https://mlanthology.org/icml/2020/jain2020icml-generalization/)BibTeX
@inproceedings{jain2020icml-generalization,
title = {{Generalization to New Actions in Reinforcement Learning}},
author = {Jain, Ayush and Szot, Andrew and Lim, Joseph},
booktitle = {International Conference on Machine Learning},
year = {2020},
pages = {4661-4672},
volume = {119},
url = {https://mlanthology.org/icml/2020/jain2020icml-generalization/}
}