Learning Visual Engagement for Trauma Recovery
Abstract
Applications ranging from human emotion understanding to e-health are exploring methods to effectively understand user behavior from self-reported questionnaires. However, little is understood about non-invasive techniques that involve face-based deep-learning models to predict engagement. Current research in visual engagement poses two key questions: 1) how much time do we need to analyze facial behavior for accurate engagement prediction? and 2) which deep learning approach provides the most accurate predictions? In this paper we compare RNN, GRU and LSTM using different length segments of AUs. Our experiments show no significant difference in prediction accuracy when using anywhere between 15 and 90 seconds of data. Moreover, the results reveal that simpler models of recurrent networks are statistically significantly better suited for capturing engagement from AUs.
Cite
Text
Dhamija and Boult. "Learning Visual Engagement for Trauma Recovery." IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 2018. doi:10.1109/WACVW.2018.00016Markdown
[Dhamija and Boult. "Learning Visual Engagement for Trauma Recovery." IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 2018.](https://mlanthology.org/wacvw/2018/dhamija2018wacvw-learning/) doi:10.1109/WACVW.2018.00016BibTeX
@inproceedings{dhamija2018wacvw-learning,
title = {{Learning Visual Engagement for Trauma Recovery}},
author = {Dhamija, Svati and Boult, Terrance E.},
booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision Workshops},
year = {2018},
pages = {84-93},
doi = {10.1109/WACVW.2018.00016},
url = {https://mlanthology.org/wacvw/2018/dhamija2018wacvw-learning/}
}