VideoMatch: Matching Based Video Object Segmentation
Abstract
Video object segmentation is challenging yet important in a wide variety of applications for video analysis. Recent works formulate video object segmentation as a prediction task using deep nets to achieve appealing state-of-the-art performance. Due to the formulation as a prediction task, most of these methods require fine-tuning during test time, such that the deep nets memorize the appearance of the objects of interest in the given video. However, fine-tuning is time-consuming and computationally expensive, hence the algorithms are far from real time. To address this issue, we develop a novel matching based algorithm for video object segmentation. In contrast to memorization based classification techniques, the proposed approach learns to match extracted features to a provided template without memorizing the appearance of the objects. We validate the effectiveness and the robustness of the proposed method on the challenging DAVIS-2016, DAVIS-2017, Youtube-Objects and JumpCut datasets. Extensive results show that our method achieves comparable performance without fine-tuning and is much more favorable in terms of computational time.
Cite
Text
Hu et al. "VideoMatch: Matching Based Video Object Segmentation." Proceedings of the European Conference on Computer Vision (ECCV), 2018. doi:10.1007/978-3-030-01237-3_4Markdown
[Hu et al. "VideoMatch: Matching Based Video Object Segmentation." Proceedings of the European Conference on Computer Vision (ECCV), 2018.](https://mlanthology.org/eccv/2018/hu2018eccv-videomatch/) doi:10.1007/978-3-030-01237-3_4BibTeX
@inproceedings{hu2018eccv-videomatch,
title = {{VideoMatch: Matching Based Video Object Segmentation}},
author = {Hu, Yuan-Ting and Huang, Jia-Bin and Schwing, Alexander G.},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2018},
doi = {10.1007/978-3-030-01237-3_4},
url = {https://mlanthology.org/eccv/2018/hu2018eccv-videomatch/}
}