Imitation Learning from Pixel Observations for Continuous Control
Abstract
We study imitation learning from visual observations only for controlling dynamical systems with continuous states and actions. This setting is attractive due to the large amount of video data available from which agents could learn from. However, it is challenging due to $i)$ not observing the actions and $ii)$ the high-dimensional visual space. In this setting, we explore recipes for imitation learning based on adversarial learning and optimal transport. These recipes enable us to scale these methods to attain expert-level performance on visual continuous control tasks in the DeepMind control suite. We investigate the tradeoffs of these approaches and present a comprehensive evaluation of the key design choices. To encourage reproducible research in this area, we provide an easy-to-use implementation for benchmarking visual imitation learning, including our methods and expert demonstrations.
Cite
Text
Cohen et al. "Imitation Learning from Pixel Observations for Continuous Control." NeurIPS 2021 Workshops: DeepRL, 2021.Markdown
[Cohen et al. "Imitation Learning from Pixel Observations for Continuous Control." NeurIPS 2021 Workshops: DeepRL, 2021.](https://mlanthology.org/neuripsw/2021/cohen2021neuripsw-imitation/)BibTeX
@inproceedings{cohen2021neuripsw-imitation,
title = {{Imitation Learning from Pixel Observations for Continuous Control}},
author = {Cohen, Samuel and Amos, Brandon and Deisenroth, Marc Peter and Henaff, Mikael and Vinitsky, Eugene and Yarats, Denis},
booktitle = {NeurIPS 2021 Workshops: DeepRL},
year = {2021},
url = {https://mlanthology.org/neuripsw/2021/cohen2021neuripsw-imitation/}
}