TriDepth: Triangular Patch-Based Deep Depth Prediction

Abstract

We propose a novel and efficient representation for single-view depth estimation using Convolutional Neural Networks (CNNs). Point-cloud is generally used for CNN-based 3D scene reconstruction; however it has some drawbacks: (1) it is redundant as a representation for planar surfaces, and (2) no spatial relationships between points are available (e.g, texture and surface). As a more efficient representation, we introduce a triangular-patch-cloud, which represents the surface of the 3D structure using a set of triangular patches, and propose a CNN framework for its 3D structure estimation. In our framework, we create it by separating all the faces in a 2D mesh, which are determined adaptively from the input image, and estimate depths and normals of all the faces. Using a common RGBD-dataset, we show that our representation has a better or comparable performance than the existing point-cloud-based methods, although it has much less parameters.

Cite

Text

Kaneko et al. "TriDepth: Triangular Patch-Based Deep Depth Prediction." IEEE/CVF International Conference on Computer Vision Workshops, 2019. doi:10.1109/ICCVW.2019.00466

Markdown

[Kaneko et al. "TriDepth: Triangular Patch-Based Deep Depth Prediction." IEEE/CVF International Conference on Computer Vision Workshops, 2019.](https://mlanthology.org/iccvw/2019/kaneko2019iccvw-tridepth/) doi:10.1109/ICCVW.2019.00466

BibTeX

@inproceedings{kaneko2019iccvw-tridepth,
  title     = {{TriDepth: Triangular Patch-Based Deep Depth Prediction}},
  author    = {Kaneko, Masaya and Sakurada, Ken and Aizawa, Kiyoharu},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2019},
  pages     = {3747-3750},
  doi       = {10.1109/ICCVW.2019.00466},
  url       = {https://mlanthology.org/iccvw/2019/kaneko2019iccvw-tridepth/}
}