Video-Based Face Alignment with Local Motion Modeling

Abstract

Face alignment remains difficult under uncontrolled conditions due to the many variations that may considerably impact facial appearance. Recently, video-based approaches have been proposed, which take advantage of temporal coherence to improve robustness. These new approaches suffer from limited temporal connectivity. We show that early, direct pixel connectivity enables the detection of local motion patterns and the learning of a hierarchy of motion features. We integrate local motion to the two predominant models in the literature, coordinate regression networks and heatmap regression networks, and combine it with late connectivity based on recurrent neural networks. The experimental results on two datasets, 300VW and SNaP-2DFe, show that local motion improves video-based face alignment and is complementary to late temporal information. Despite the simplicity of the proposed architectures, our best model provides competitive performance with more complex models from the literature.

Cite

Text

Belmonte et al. "Video-Based Face Alignment with Local Motion Modeling." IEEE/CVF Winter Conference on Applications of Computer Vision, 2019. doi:10.1109/WACV.2019.00228

Markdown

[Belmonte et al. "Video-Based Face Alignment with Local Motion Modeling." IEEE/CVF Winter Conference on Applications of Computer Vision, 2019.](https://mlanthology.org/wacv/2019/belmonte2019wacv-video/) doi:10.1109/WACV.2019.00228

BibTeX

@inproceedings{belmonte2019wacv-video,
  title     = {{Video-Based Face Alignment with Local Motion Modeling}},
  author    = {Belmonte, Romain and Ihaddadene, Nacim and Tirilly, Pierre and Bilasco, Ioan Marius and Djeraba, Chaabane},
  booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision},
  year      = {2019},
  pages     = {2106-2115},
  doi       = {10.1109/WACV.2019.00228},
  url       = {https://mlanthology.org/wacv/2019/belmonte2019wacv-video/}
}