Facial Action Unit Event Detection by Cascade of Tasks
Abstract
Automatic facial Action Unit (AU) detection from video is a long-standing problem in facial expression analysis. AU detection is typically posed as a classification problem between frames or segments of positive examples and negative ones, where existing work emphasizes the use of different features or classifiers. In this paper, we propose a method called Cascade of Tasks (CoT) that combines the use of different tasks (i.e., frame, segment and transition) for AU event detection. We train CoT in a sequential manner embracing diversity, which ensures robustness and generalization to unseen data. In addition to conventional framebased metrics that evaluate frames independently, we propose a new event-based metric to evaluate detection performance at event-level. We show how the CoT method consistently outperforms state-of-the-art approaches in both frame-based and event-based metrics, across three public datasets that differ in complexity: CK+, FERA and RUFACS.
Cite
Text
Ding et al. "Facial Action Unit Event Detection by Cascade of Tasks." International Conference on Computer Vision, 2013. doi:10.1109/ICCV.2013.298Markdown
[Ding et al. "Facial Action Unit Event Detection by Cascade of Tasks." International Conference on Computer Vision, 2013.](https://mlanthology.org/iccv/2013/ding2013iccv-facial/) doi:10.1109/ICCV.2013.298BibTeX
@inproceedings{ding2013iccv-facial,
title = {{Facial Action Unit Event Detection by Cascade of Tasks}},
author = {Ding, Xiaoyu and Chu, Wen-Sheng and De La Torre, Fernando and Cohn, Jeffery F. and Wang, Qiao},
booktitle = {International Conference on Computer Vision},
year = {2013},
doi = {10.1109/ICCV.2013.298},
url = {https://mlanthology.org/iccv/2013/ding2013iccv-facial/}
}