Action-Conditioned Convolutional Future Regression Models for Robot Imitation Learning

Abstract

Based on what is seen (i.e. visual input), humans are able to visually predict (i.e. regress) what the scene will look like after taking a certain action. Further, humans are able to take advantage of such predictions to select optimal actions for the task they are working on. Using example videos, robots can also learn to visually imagine the future consequence of taking an action. This can be viewed as learning a function mapping a raw image frame (conditioned on a particular action) to the future image frame. Once learned, the future regression function can be combined with an action policy learning framework (e.g. reinforcement or imitation learning), enabling better robot action learning for given tasks.

Cite

Text

Wu et al. "Action-Conditioned Convolutional Future Regression Models for Robot Imitation Learning." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018. doi:10.1109/CVPRW.2018.00274

Markdown

[Wu et al. "Action-Conditioned Convolutional Future Regression Models for Robot Imitation Learning." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2018.](https://mlanthology.org/cvprw/2018/wu2018cvprw-actionconditioned/) doi:10.1109/CVPRW.2018.00274

BibTeX

@inproceedings{wu2018cvprw-actionconditioned,
  title     = {{Action-Conditioned Convolutional Future Regression Models for Robot Imitation Learning}},
  author    = {Wu, Alan and Piergiovanni, A. J. and Ryoo, Michael S.},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2018},
  pages     = {2035-2037},
  doi       = {10.1109/CVPRW.2018.00274},
  url       = {https://mlanthology.org/cvprw/2018/wu2018cvprw-actionconditioned/}
}