A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis

Abstract

Video segmentation research is currently limited by the lack of a benchmark dataset that covers the large variety of subproblems appearing in video segmentation and that is large enough to avoid overfitting. Consequently, there is little analysis of video segmentation which generalizes across subtasks, and it is not yet clear which and how video segmentation should leverage the information from the still-frames, as previously studied in image segmentation, alongside video specific information, such as temporal volume, motion and occlusion. In this work we provide such an analysis based on annotations of a large video dataset, where each video is manually segmented by multiple persons. Moreover, we introduce a new volume-based metric that includes the important aspect of temporal consistency, that can deal with segmentation hierarchies, and that reflects the tradeoff between over-segmentation and segmentation accuracy.

Cite

Text

Galasso et al. "A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis." International Conference on Computer Vision, 2013. doi:10.1109/ICCV.2013.438

Markdown

[Galasso et al. "A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis." International Conference on Computer Vision, 2013.](https://mlanthology.org/iccv/2013/galasso2013iccv-unified/) doi:10.1109/ICCV.2013.438

BibTeX

@inproceedings{galasso2013iccv-unified,
  title     = {{A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis}},
  author    = {Galasso, Fabio and Nagaraja, Naveen Shankar and Cardenas, Tatiana Jimenez and Brox, Thomas and Schiele, Bernt},
  booktitle = {International Conference on Computer Vision},
  year      = {2013},
  doi       = {10.1109/ICCV.2013.438},
  url       = {https://mlanthology.org/iccv/2013/galasso2013iccv-unified/}
}