Addressing Sample Complexity in Visual Tasks Using Hindsight Experience Replay and Hallucinatory GANs

Abstract

Reinforcement Learning (RL) algorithms typically require millions of environment interactions to learn successful policies in sparse reward settings. Hindsight Experience Replay (HER) was introduced as a technique to increase sample efficiency by re-imagining unsuccessful trajectories as successful ones by changing the originally intended goals. However, HER cannot be directly applied to visual environments where goal states are characterized by the presence of distinct visual features. In this work, we show how visual trajectories can be hallucinated to appear successful by altering agent observations using a generative model trained on relatively few snapshots of the goal. We then use this model in combination with HER to train RL agents in visual settings. We validate our approach on 3D navigation tasks and a simulated robotics application and show marked improvement over standard RL algorithms and baselines derived from previous work.

Cite

Text

Sahni et al. "Addressing Sample Complexity in Visual Tasks Using Hindsight Experience Replay and Hallucinatory GANs." ICML 2019 Workshops: RL4RealLife, 2019.

Markdown

[Sahni et al. "Addressing Sample Complexity in Visual Tasks Using Hindsight Experience Replay and Hallucinatory GANs." ICML 2019 Workshops: RL4RealLife, 2019.](https://mlanthology.org/icmlw/2019/sahni2019icmlw-addressing/)

BibTeX

@inproceedings{sahni2019icmlw-addressing,
  title     = {{Addressing Sample Complexity in Visual Tasks Using Hindsight Experience Replay and Hallucinatory GANs}},
  author    = {Sahni, Himanshu and Buckley, Toby and Abbeel, Pieter and Kuzovkin, Ilya},
  booktitle = {ICML 2019 Workshops: RL4RealLife},
  year      = {2019},
  url       = {https://mlanthology.org/icmlw/2019/sahni2019icmlw-addressing/}
}