Recurrent Fully Convolutional Networks for Video Segmentation

Abstract

Image segmentation is an important step in most visual tasks. While convolutional neural networks have shown to perform well on single image segmentation, to our knowledge, no study has been done on leveraging recurrent gated architectures for video segmentation. Accordingly, we propose and implement a novel method for online segmentation of video sequences that incorporates temporal data. The network is built from a fully convolutional network and a recurrent unit that works on a sliding window over the temporal data. We use convolutional gated recurrent unit that preserves the spatial information and reduces the parameters learned. Our method has the advantage that it can work in an online fashion instead of operating over the whole input batch of video frames. The network is tested on video segmentation benchmarks in Segtrack V2 and Davis. It proved to have 5% improvement in Segtrack and 3% improvement in Davis in F-measure over a plain fully convolutional network.

Cite

Text

Valipour et al. "Recurrent Fully Convolutional Networks for Video Segmentation." IEEE/CVF Winter Conference on Applications of Computer Vision, 2017. doi:10.1109/WACV.2017.11

Markdown

[Valipour et al. "Recurrent Fully Convolutional Networks for Video Segmentation." IEEE/CVF Winter Conference on Applications of Computer Vision, 2017.](https://mlanthology.org/wacv/2017/valipour2017wacv-recurrent/) doi:10.1109/WACV.2017.11

BibTeX

@inproceedings{valipour2017wacv-recurrent,
  title     = {{Recurrent Fully Convolutional Networks for Video Segmentation}},
  author    = {Valipour, Sepehr and Siam, Mennatullah and Jägersand, Martin and Ray, Nilanjan},
  booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision},
  year      = {2017},
  pages     = {29-36},
  doi       = {10.1109/WACV.2017.11},
  url       = {https://mlanthology.org/wacv/2017/valipour2017wacv-recurrent/}
}