SLAM3R: Real-Time Dense Scene Reconstruction from Monocular RGB Videos

Abstract

In this paper, we introduce SLAM3R, a novel and effective system for real-time, high-quality, dense 3D reconstruction using RGB videos. SLAM3R provides an end-to-end solution by seamlessly integrating local 3D reconstruction and global coordinate registration through feed-forward neural networks. Given an input video, the system first converts it into overlapping clips using a sliding window mechanism. Unlike traditional pose optimization-based methods, SLAM3R directly regresses 3D pointmaps from RGB images in each window and progressively aligns and deforms these local pointmaps to create a globally consistent scene reconstruction - all without explicitly solving any camera parameters. Experiments across datasets consistently show that SLAM3R achieves state-of-the-art reconstruction accuracy and completeness while maintaining real-time performance at 20+ FPS. Code available at: https://github.com/PKU-VCL-3DV/SLAM3R.

Cite

Text

Liu et al. "SLAM3R: Real-Time Dense Scene Reconstruction from Monocular RGB Videos." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.01552

Markdown

[Liu et al. "SLAM3R: Real-Time Dense Scene Reconstruction from Monocular RGB Videos." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/liu2025cvpr-slam3r/) doi:10.1109/CVPR52734.2025.01552

BibTeX

@inproceedings{liu2025cvpr-slam3r,
  title     = {{SLAM3R: Real-Time Dense Scene Reconstruction from Monocular RGB Videos}},
  author    = {Liu, Yuzheng and Dong, Siyan and Wang, Shuzhe and Yin, Yingda and Yang, Yanchao and Fan, Qingnan and Chen, Baoquan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {16651-16662},
  doi       = {10.1109/CVPR52734.2025.01552},
  url       = {https://mlanthology.org/cvpr/2025/liu2025cvpr-slam3r/}
}