Dense Monocular Depth Estimation in Complex Dynamic Scenes
Abstract
We present an approach to dense depth estimation from a single monocular camera that is moving through a dynamic scene. The approach produces a dense depth map from two consecutive frames. Moving objects are reconstructed along with the surrounding environment. We provide a novel motion segmentation algorithm that segments the optical flow field into a set of motion models, each with its own epipolar geometry. We then show that the scene can be reconstructed based on these motion models by optimizing a convex program. The optimization jointly reasons about the scales of different objects and assembles the scene in a common coordinate frame, determined up to a global scale. Experimental results demonstrate that the presented approach outperforms prior methods for monocular depth estimation in dynamic scenes.
Cite
Text
Ranftl et al. "Dense Monocular Depth Estimation in Complex Dynamic Scenes." Conference on Computer Vision and Pattern Recognition, 2016. doi:10.1109/CVPR.2016.440Markdown
[Ranftl et al. "Dense Monocular Depth Estimation in Complex Dynamic Scenes." Conference on Computer Vision and Pattern Recognition, 2016.](https://mlanthology.org/cvpr/2016/ranftl2016cvpr-dense/) doi:10.1109/CVPR.2016.440BibTeX
@inproceedings{ranftl2016cvpr-dense,
title = {{Dense Monocular Depth Estimation in Complex Dynamic Scenes}},
author = {Ranftl, Rene and Vineet, Vibhav and Chen, Qifeng and Koltun, Vladlen},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2016},
doi = {10.1109/CVPR.2016.440},
url = {https://mlanthology.org/cvpr/2016/ranftl2016cvpr-dense/}
}