A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis
Abstract
Video segmentation research is currently limited by the lack of a benchmark dataset that covers the large variety of subproblems appearing in video segmentation and that is large enough to avoid overfitting. Consequently, there is little analysis of video segmentation which generalizes across subtasks, and it is not yet clear which and how video segmentation should leverage the information from the still-frames, as previously studied in image segmentation, alongside video specific information, such as temporal volume, motion and occlusion. In this work we provide such an analysis based on annotations of a large video dataset, where each video is manually segmented by multiple persons. Moreover, we introduce a new volume-based metric that includes the important aspect of temporal consistency, that can deal with segmentation hierarchies, and that reflects the tradeoff between over-segmentation and segmentation accuracy.
Cite
Text
Galasso et al. "A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis." International Conference on Computer Vision, 2013. doi:10.1109/ICCV.2013.438Markdown
[Galasso et al. "A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis." International Conference on Computer Vision, 2013.](https://mlanthology.org/iccv/2013/galasso2013iccv-unified/) doi:10.1109/ICCV.2013.438BibTeX
@inproceedings{galasso2013iccv-unified,
title = {{A Unified Video Segmentation Benchmark: Annotation, Metrics and Analysis}},
author = {Galasso, Fabio and Nagaraja, Naveen Shankar and Cardenas, Tatiana Jimenez and Brox, Thomas and Schiele, Bernt},
booktitle = {International Conference on Computer Vision},
year = {2013},
doi = {10.1109/ICCV.2013.438},
url = {https://mlanthology.org/iccv/2013/galasso2013iccv-unified/}
}