Dense Monocular Depth Estimation in Complex Dynamic Scenes

Abstract

We present an approach to dense depth estimation from a single monocular camera that is moving through a dynamic scene. The approach produces a dense depth map from two consecutive frames. Moving objects are reconstructed along with the surrounding environment. We provide a novel motion segmentation algorithm that segments the optical flow field into a set of motion models, each with its own epipolar geometry. We then show that the scene can be reconstructed based on these motion models by optimizing a convex program. The optimization jointly reasons about the scales of different objects and assembles the scene in a common coordinate frame, determined up to a global scale. Experimental results demonstrate that the presented approach outperforms prior methods for monocular depth estimation in dynamic scenes.

Cite

Text

Ranftl et al. "Dense Monocular Depth Estimation in Complex Dynamic Scenes." Conference on Computer Vision and Pattern Recognition, 2016. doi:10.1109/CVPR.2016.440

Markdown

[Ranftl et al. "Dense Monocular Depth Estimation in Complex Dynamic Scenes." Conference on Computer Vision and Pattern Recognition, 2016.](https://mlanthology.org/cvpr/2016/ranftl2016cvpr-dense/) doi:10.1109/CVPR.2016.440

BibTeX

@inproceedings{ranftl2016cvpr-dense,
  title     = {{Dense Monocular Depth Estimation in Complex Dynamic Scenes}},
  author    = {Ranftl, Rene and Vineet, Vibhav and Chen, Qifeng and Koltun, Vladlen},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2016},
  doi       = {10.1109/CVPR.2016.440},
  url       = {https://mlanthology.org/cvpr/2016/ranftl2016cvpr-dense/}
}