The Ingredients of Real World Robotic Reinforcement Learning

Abstract

The success of reinforcement learning in the real world has been limited to instrumented laboratory scenarios, often requiring arduous human supervision to enable continuous learning. In this work, we discuss the required elements of a robotic system that can continually and autonomously improve with data collected in the real world, and propose a particular instantiation of such a system. Subsequently, we investigate a number of challenges of learning without instrumentation -- including the lack of episodic resets, state estimation, and hand-engineered rewards -- and propose simple, scalable solutions to these challenges. We demonstrate the efficacy of our proposed system on dexterous robotic manipulation tasks in simulation and the real world, and also provide an insightful analysis and ablation study of the challenges associated with this learning paradigm.

Cite

Text

Zhu et al. "The Ingredients of Real World Robotic Reinforcement Learning." International Conference on Learning Representations, 2020.

Markdown

[Zhu et al. "The Ingredients of Real World Robotic Reinforcement Learning." International Conference on Learning Representations, 2020.](https://mlanthology.org/iclr/2020/zhu2020iclr-ingredients/)

BibTeX

@inproceedings{zhu2020iclr-ingredients,
  title     = {{The Ingredients of Real World Robotic Reinforcement Learning}},
  author    = {Zhu, Henry and Yu, Justin and Gupta, Abhishek and Shah, Dhruv and Hartikainen, Kristian and Singh, Avi and Kumar, Vikash and Levine, Sergey},
  booktitle = {International Conference on Learning Representations},
  year      = {2020},
  url       = {https://mlanthology.org/iclr/2020/zhu2020iclr-ingredients/}
}