Associate Latent Encodings in Learning from Demonstrations

Abstract

We contribute a learning from demonstration approach for robots to acquire skills from multi-modal high-dimensional data. Both latent representations and associations of different modalities are proposed to be jointly learned through an adapted variational auto-encoder. The implementation and results are demonstrated in a robotic handwriting scenario, where the visual sensory input and the arm joint writing motion are learned and coupled. We show the latent representations successfully construct a task manifold for the observed sensor modalities. Moreover, the learned associations can be exploited to directly synthesize arm joint handwriting motion from an image input in an end-to-end manner. The advantages of learning associative latent encodings are further highlighted with the examples of inferring upon incomplete input images. A comparison with alternative methods demonstrates the superiority of the present approach in these challenging tasks.

Cite

Text

Yin et al. "Associate Latent Encodings in Learning from Demonstrations." AAAI Conference on Artificial Intelligence, 2017. doi:10.1609/AAAI.V31I1.11040

Markdown

[Yin et al. "Associate Latent Encodings in Learning from Demonstrations." AAAI Conference on Artificial Intelligence, 2017.](https://mlanthology.org/aaai/2017/yin2017aaai-associate/) doi:10.1609/AAAI.V31I1.11040

BibTeX

@inproceedings{yin2017aaai-associate,
  title     = {{Associate Latent Encodings in Learning from Demonstrations}},
  author    = {Yin, Hang and Melo, Francisco S. and Billard, Aude and Paiva, Ana},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2017},
  pages     = {3848-3854},
  doi       = {10.1609/AAAI.V31I1.11040},
  url       = {https://mlanthology.org/aaai/2017/yin2017aaai-associate/}
}