Tracking Benchmark Databases for Video-Based Sign Language Recognition

Abstract

A survey of video databases that can be used within a continuous sign language recognition scenario to measure the performance of head and hand tracking algorithms either w.r.t. a tracking error rate or w.r.t. a word error rate criterion is presented in this work. Robust tracking algorithms are required as the signing hand frequently moves in front of the face, may temporarily disappear, or cross the other hand. Only few studies consider the recognition of continuous sign language, and usually special devices such as colored gloves or blue-boxing environments are used to accurately track the regions-of-interest in sign language processing. Ground-truth labels for hand and head positions have been annotated for more than 30k frames in several publicly available video databases of different degrees of difficulty, and preliminary tracking results are presented.

Cite

Text

Dreuw et al. "Tracking Benchmark Databases for Video-Based Sign Language Recognition." European Conference on Computer Vision, 2010. doi:10.1007/978-3-642-35749-7_22

Markdown

[Dreuw et al. "Tracking Benchmark Databases for Video-Based Sign Language Recognition." European Conference on Computer Vision, 2010.](https://mlanthology.org/eccv/2010/dreuw2010eccv-tracking/) doi:10.1007/978-3-642-35749-7_22

BibTeX

@inproceedings{dreuw2010eccv-tracking,
  title     = {{Tracking Benchmark Databases for Video-Based Sign Language Recognition}},
  author    = {Dreuw, Philippe and Forster, Jens and Ney, Hermann},
  booktitle = {European Conference on Computer Vision},
  year      = {2010},
  pages     = {286-297},
  doi       = {10.1007/978-3-642-35749-7_22},
  url       = {https://mlanthology.org/eccv/2010/dreuw2010eccv-tracking/}
}