UnsupervisedR&R: Unsupervised Point Cloud Registration via Differentiable Rendering

Abstract

Aligning partial views of a scene into a single whole is essential to understanding one's environment and is a key component of numerous robotics tasks such as SLAM and SfM. Recent approaches have proposed end-to-end systems that can outperform traditional methods by leveraging pose supervision. However, with the rising prevalence of cameras with depth sensors, we can expect a new stream of raw RGB-D data without the annotations needed for supervision. We propose UnsupervisedR&R: an end-to-end unsupervised approach to learning point cloud registration from raw RGB-D video. The key idea is to leverage differentiable alignment and rendering to enforce photometric and geometric consistency between frames. We evaluate our approach on indoor scene datasets and find that we outperform existing traditional approaches with classical and learned descriptors while being competitive with supervised geometric point cloud registration approaches.

Cite

Text

El Banani et al. "UnsupervisedR&R: Unsupervised Point Cloud Registration via Differentiable Rendering." Conference on Computer Vision and Pattern Recognition, 2021.

Markdown

[El Banani et al. "UnsupervisedR&R: Unsupervised Point Cloud Registration via Differentiable Rendering." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/banani2021cvpr-unsupervisedr/)

BibTeX

@inproceedings{banani2021cvpr-unsupervisedr,
  title     = {{UnsupervisedR&R: Unsupervised Point Cloud Registration via Differentiable Rendering}},
  author    = {El Banani, Mohamed and Gao, Luya and Johnson, Justin},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2021},
  pages     = {7129-7139},
  url       = {https://mlanthology.org/cvpr/2021/banani2021cvpr-unsupervisedr/}
}