Person Verification by Lip-Motion

Abstract

This paper describes a new motion based feature extraction technique for speaker recognition using orientation estimation in 2D manifolds. The motion is estimated by computing the components of the structure tensor from which normal flows are extracted. By projecting the 3D spatiotemporal data to 2-D planes we obtain projection coefficients which we use to evaluate the 3-D orientations of brightness patterns in TV like 2D image sequences. This corresponds to the solutions of simple matrix eigenvalue problems in 2D, affording increased computational efficiency. An implementation based on joint lip movements and speech is presented along with experiments which confirm the theory, exhibiting a recognition rate of 98% on the publicly available XM2VTS database

Cite

Text

Faraj and Bigün. "Person Verification by Lip-Motion." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2006. doi:10.1109/CVPRW.2006.158

Markdown

[Faraj and Bigün. "Person Verification by Lip-Motion." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2006.](https://mlanthology.org/cvprw/2006/faraj2006cvprw-person/) doi:10.1109/CVPRW.2006.158

BibTeX

@inproceedings{faraj2006cvprw-person,
  title     = {{Person Verification by Lip-Motion}},
  author    = {Faraj, Maycel Isaac and Bigün, Josef},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2006},
  pages     = {37},
  doi       = {10.1109/CVPRW.2006.158},
  url       = {https://mlanthology.org/cvprw/2006/faraj2006cvprw-person/}
}