Learning to Look by Self-Prediction

Abstract

We present a method for learning active vision skills, to move the camera to observe a robot's sensors from informative points of view, without external rewards or labels. We do this by jointly training a visual predictor network, which predicts future returns of the sensors using pixels, and a camera control agent, which we reward using the negative error of the predictor. The agent thus moves the camera to points of view that are most predictive for a chosen sensor, which we select using a conditioning input to the agent. We observe that despite this noisy learned reward function, the learned policies a exhibit competence by reliably framing the sensor in a specific location in the view, an emergent location which we call a behavioral fovea. We find that replacing the conventional camera with a foveal camera further increases the policies' precision.

Cite

Text

Grimes et al. "Learning to Look by Self-Prediction." Transactions on Machine Learning Research, 2023.

Markdown

[Grimes et al. "Learning to Look by Self-Prediction." Transactions on Machine Learning Research, 2023.](https://mlanthology.org/tmlr/2023/grimes2023tmlr-learning/)

BibTeX

@article{grimes2023tmlr-learning,
  title     = {{Learning to Look by Self-Prediction}},
  author    = {Grimes, Matthew Koichi and Modayil, Joseph Varughese and Mirowski, Piotr W and Rao, Dushyant and Hadsell, Raia},
  journal   = {Transactions on Machine Learning Research},
  year      = {2023},
  url       = {https://mlanthology.org/tmlr/2023/grimes2023tmlr-learning/}
}