Human-Assisted Motion Annotation

Abstract

Obtaining ground-truth motion for arbitrary, real-world video sequences is a challenging but important task for both algorithm evaluation and model design. Existing ground-truth databases are either synthetic, such as the Yosemite sequence, or limited to indoor, experimental setups, such as the database developed by Baker et al (2007). We propose a human-in-loop methodology to create a ground-truth motion database for the videos taken with ordinary cameras in both indoor and outdoor scenes, using the fact that human beings are experts at segmenting objects and inspecting the match between two frames. We designed an interactive computer vision system to allow a user to efficiently annotate motion. Our methodology is cross-validated by showing that human annotated motion is repeatable, consistent across annotators, and close to the ground truth obtained by Baker et al (2007). Using our system, we collected and annotated 10 indoor and outdoor real-world videos to form a ground-truth motion database. The source code, annotation tool and database is online for public evaluation and benchmarking.

Cite

Text

Liu et al. "Human-Assisted Motion Annotation." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2008. doi:10.1109/CVPR.2008.4587845

Markdown

[Liu et al. "Human-Assisted Motion Annotation." IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2008.](https://mlanthology.org/cvpr/2008/liu2008cvpr-human/) doi:10.1109/CVPR.2008.4587845

BibTeX

@inproceedings{liu2008cvpr-human,
  title     = {{Human-Assisted Motion Annotation}},
  author    = {Liu, Ce and Freeman, William T. and Adelson, Edward H. and Weiss, Yair},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year      = {2008},
  doi       = {10.1109/CVPR.2008.4587845},
  url       = {https://mlanthology.org/cvpr/2008/liu2008cvpr-human/}
}