Accurate, Real-Time, Unadorned Lip Tracking
Abstract
Human speech is inherently multi-modal, consisting of both audio and visual components. Recently researchers have shown that the incorporation of information about the position of the lips into acoustic speech recognisers enables robust recognition of noisy speech. In the case of Hidden Markov Model-recognition, we show that this happens because the visual signal stabilises the alignment of states. It is also shown, that unadorned lips, both the inner and outer contours, can be robustly tracked in real time on general-purpose workstations. To accomplish this, efficient algorithms are employed which contain three key components: shape models, motion models, and focused colour feature detectors-all of which are learnt from examples.
Cite
Text
Kaucic and Blake. "Accurate, Real-Time, Unadorned Lip Tracking." IEEE/CVF International Conference on Computer Vision, 1998. doi:10.1109/ICCV.1998.710745Markdown
[Kaucic and Blake. "Accurate, Real-Time, Unadorned Lip Tracking." IEEE/CVF International Conference on Computer Vision, 1998.](https://mlanthology.org/iccv/1998/kaucic1998iccv-accurate/) doi:10.1109/ICCV.1998.710745BibTeX
@inproceedings{kaucic1998iccv-accurate,
title = {{Accurate, Real-Time, Unadorned Lip Tracking}},
author = {Kaucic, Robert and Blake, Andrew},
booktitle = {IEEE/CVF International Conference on Computer Vision},
year = {1998},
pages = {370-375},
doi = {10.1109/ICCV.1998.710745},
url = {https://mlanthology.org/iccv/1998/kaucic1998iccv-accurate/}
}