Semantic Segmentation of RGBD Videos with Recurrent Fully Convolutional Neural Networks

Abstract

Semantic segmentation of videos using neural networks is currently a popular task, the work done in this field is however mostly on RGB videos. The main reason for this is the lack of large RGBD video datasets, annotated with ground truth information at the pixel level. In this work, we use a synthetic RGBD video dataset to investigate the contribution of depth and temporal information to the video segmentation task using convolutional and recurrent neural network architectures. Our experiments show the addition of depth information improves semantic segmentation results and exploiting temporal information results in higher quality output segmentations.

Cite

Text

Yurdakul and Yemez. "Semantic Segmentation of RGBD Videos with Recurrent Fully Convolutional Neural Networks." IEEE/CVF International Conference on Computer Vision Workshops, 2017. doi:10.1109/ICCVW.2017.51

Markdown

[Yurdakul and Yemez. "Semantic Segmentation of RGBD Videos with Recurrent Fully Convolutional Neural Networks." IEEE/CVF International Conference on Computer Vision Workshops, 2017.](https://mlanthology.org/iccvw/2017/yurdakul2017iccvw-semantic/) doi:10.1109/ICCVW.2017.51

BibTeX

@inproceedings{yurdakul2017iccvw-semantic,
  title     = {{Semantic Segmentation of RGBD Videos with Recurrent Fully Convolutional Neural Networks}},
  author    = {Yurdakul, Ekrem Emre and Yemez, Yucel},
  booktitle = {IEEE/CVF International Conference on Computer Vision Workshops},
  year      = {2017},
  pages     = {367-374},
  doi       = {10.1109/ICCVW.2017.51},
  url       = {https://mlanthology.org/iccvw/2017/yurdakul2017iccvw-semantic/}
}