Depth mAP Prediction from a Single Image Using a Multi-Scale Deep Network

Abstract

Predicting depth is an essential component in understanding the 3D geometry of a scene. While for stereo images local correspondence suffices for estimation, finding depth relations from a single image is less straightforward, requiring integration of both global and local information from various cues. Moreover, the task is inherently ambiguous, with a large source of uncertainty coming from the overall scale. In this paper, we present a new method that addresses this task by employing two deep network stacks: one that makes a coarse global prediction based on the entire image, and another that refines this prediction locally. We also apply a scale-invariant error to help measure depth relations rather than scale. By leveraging the raw datasets as large sources of training data, our method achieves state-of-the-art results on both NYU Depth and KITTI, and matches detailed depth boundaries without the need for superpixelation.

Cite

Text

Eigen et al. "Depth mAP Prediction from a Single Image Using a Multi-Scale Deep Network." Neural Information Processing Systems, 2014.

Markdown

[Eigen et al. "Depth mAP Prediction from a Single Image Using a Multi-Scale Deep Network." Neural Information Processing Systems, 2014.](https://mlanthology.org/neurips/2014/eigen2014neurips-depth/)

BibTeX

@inproceedings{eigen2014neurips-depth,
  title     = {{Depth mAP Prediction from a Single Image Using a Multi-Scale Deep Network}},
  author    = {Eigen, David and Puhrsch, Christian and Fergus, Rob},
  booktitle = {Neural Information Processing Systems},
  year      = {2014},
  pages     = {2366-2374},
  url       = {https://mlanthology.org/neurips/2014/eigen2014neurips-depth/}
}