Learning Common Representation from RGB and Depth Images

Abstract

We propose a new deep learning architecture for the tasks of semantic segmentation and depth prediction from RGB-D images. We revise the state of art based on the RGB and depth feature fusion, where both modalities are assumed to be available at train and test time. We propose a new architecture where the feature fusion is replaced with a common deep representation. Combined with an encoder-decoder type of the network, the architecture can jointly learn models for semantic segmentation and depth estimation based on their common representation. This representation, inspired by multi-view learning, offers several important advantages, such as using one modality available at test time to reconstruct the missing modality. In the RGB-D case, this enables the cross-modality scenarios, such as using depth data for semantically segmentation and the RGB images for depth estimation. We demonstrate the effectiveness of the proposed network on two publicly available RGB-D datasets. The experimental results show that the proposed method works well in both semantic segmentation and depth estimation tasks.

Cite

Text

Giannone and Chidlovskii. "Learning Common Representation from RGB and Depth Images." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019. doi:10.1109/CVPRW.2019.00054

Markdown

[Giannone and Chidlovskii. "Learning Common Representation from RGB and Depth Images." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2019.](https://mlanthology.org/cvprw/2019/giannone2019cvprw-learning/) doi:10.1109/CVPRW.2019.00054

BibTeX

@inproceedings{giannone2019cvprw-learning,
  title     = {{Learning Common Representation from RGB and Depth Images}},
  author    = {Giannone, Giorgio and Chidlovskii, Boris},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2019},
  pages     = {408-415},
  doi       = {10.1109/CVPRW.2019.00054},
  url       = {https://mlanthology.org/cvprw/2019/giannone2019cvprw-learning/}
}