VIRDO++: Real-World, Visuo-Tactile Dynamics and Perception of Deformable Objects

Abstract

Deformable objects manipulation can benefit from representations that seamlessly integrate vision and touch while handling occlusions. In this work, we present a novel approach for, and real-world demonstration of, multimodal visuo-tactile state-estimation and dynamics prediction for deformable objects. Our approach, VIRDO++, builds on recent progress in multimodal neural implicit representations for deformable object state-estimation (VIRDO) via a new formulation for deformation dynamics and a complementary state-estimation algorithm that (i) maintains a belief over deformations, and (ii) enables practical real-world application by removing the need for privileged contact information. In the context of two real-world robotic tasks, we show: (i) high-fidelity cross-modal state-estimation and prediction of deformable objects from partial visuo-tactile feedback, and (ii) generalization to unseen objects and contact formations.

Cite

Text

Wi et al. "VIRDO++: Real-World, Visuo-Tactile Dynamics and Perception of Deformable Objects." Conference on Robot Learning, 2022.

Markdown

[Wi et al. "VIRDO++: Real-World, Visuo-Tactile Dynamics and Perception of Deformable Objects." Conference on Robot Learning, 2022.](https://mlanthology.org/corl/2022/wi2022corl-virdo/)

BibTeX

@inproceedings{wi2022corl-virdo,
  title     = {{VIRDO++: Real-World, Visuo-Tactile Dynamics and Perception of Deformable Objects}},
  author    = {Wi, Youngsun and Zeng, Andy and Florence, Pete and Fazeli, Nima},
  booktitle = {Conference on Robot Learning},
  year      = {2022},
  pages     = {1806-1816},
  volume    = {205},
  url       = {https://mlanthology.org/corl/2022/wi2022corl-virdo/}
}