Curiosity-Driven Exploration by Self-Supervised Prediction

Abstract

In many real-world scenarios, rewards extrinsic to the agent are extremely sparse, or absent altogether. In such cases, curiosity can serve as an intrinsic reward signal to enable the agent to explore its environment and learn skills that might be useful later in its life. We formulate curiosity as the error in an agent's ability to predict the consequence of its own actions in a visual feature space learned by a self-supervised inverse dynamics model. Our formulation scales to high-dimensional continuous state spaces like images, bypasses the difficulties of directly predicting pixels, and, critically, ignores the aspects of the environment that cannot affect the agent. The proposed approach is evaluated in two environments: VizDoom and Super Mario Bros. Three broad settings are investigated: 1) sparse extrinsic reward; 2) exploration with no extrinsic reward; and 3) generalization to unseen scenarios (e.g. new levels of the same game).

Cite

Text

Pathak et al. "Curiosity-Driven Exploration by Self-Supervised Prediction." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017. doi:10.1109/CVPRW.2017.70

Markdown

[Pathak et al. "Curiosity-Driven Exploration by Self-Supervised Prediction." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017.](https://mlanthology.org/cvprw/2017/pathak2017cvprw-curiositydriven/) doi:10.1109/CVPRW.2017.70

BibTeX

@inproceedings{pathak2017cvprw-curiositydriven,
  title     = {{Curiosity-Driven Exploration by Self-Supervised Prediction}},
  author    = {Pathak, Deepak and Agrawal, Pulkit and Efros, Alexei A. and Darrell, Trevor},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2017},
  pages     = {488-489},
  doi       = {10.1109/CVPRW.2017.70},
  url       = {https://mlanthology.org/cvprw/2017/pathak2017cvprw-curiositydriven/}
}