DTAM: Dense Tracking and Mapping in Real-Time

Abstract

DTAM is a system for real-time camera tracking and reconstruction which relies not on feature extraction but dense, every pixel methods. As a single hand-held RGB camera flies over a static scene, we estimate detailed textured depth maps at selected keyframes to produce a surface patchwork with millions of vertices. We use the hundreds of images available in a video stream to improve the quality of a simple photometric data term, and minimise a global spatially regularised energy functional in a novel non-convex optimisation framework. Interleaved, we track the camera's 6DOF motion precisely by frame-rate whole image alignment against the entire dense model. Our algorithms are highly parallelisable throughout and DTAM achieves real-time performance using current commodity GPU hardware. We demonstrate that a dense model permits superior tracking performance under rapid motion compared to a state of the art method using features; and also show the additional usefulness of the dense model for real-time scene interaction in a physics-enhanced augmented reality application.

Cite

Text

Newcombe et al. "DTAM: Dense Tracking and Mapping in Real-Time." IEEE/CVF International Conference on Computer Vision, 2011. doi:10.1109/ICCV.2011.6126513

Markdown

[Newcombe et al. "DTAM: Dense Tracking and Mapping in Real-Time." IEEE/CVF International Conference on Computer Vision, 2011.](https://mlanthology.org/iccv/2011/newcombe2011iccv-dtam/) doi:10.1109/ICCV.2011.6126513

BibTeX

@inproceedings{newcombe2011iccv-dtam,
  title     = {{DTAM: Dense Tracking and Mapping in Real-Time}},
  author    = {Newcombe, Richard A. and Lovegrove, Steven and Davison, Andrew J.},
  booktitle = {IEEE/CVF International Conference on Computer Vision},
  year      = {2011},
  pages     = {2320-2327},
  doi       = {10.1109/ICCV.2011.6126513},
  url       = {https://mlanthology.org/iccv/2011/newcombe2011iccv-dtam/}
}