Learning Visual Engagement for Trauma Recovery

Abstract

Applications ranging from human emotion understanding to e-health are exploring methods to effectively understand user behavior from self-reported questionnaires. However, little is understood about non-invasive techniques that involve face-based deep-learning models to predict engagement. Current research in visual engagement poses two key questions: 1) how much time do we need to analyze facial behavior for accurate engagement prediction? and 2) which deep learning approach provides the most accurate predictions? In this paper we compare RNN, GRU and LSTM using different length segments of AUs. Our experiments show no significant difference in prediction accuracy when using anywhere between 15 and 90 seconds of data. Moreover, the results reveal that simpler models of recurrent networks are statistically significantly better suited for capturing engagement from AUs.

Cite

Text

Dhamija and Boult. "Learning Visual Engagement for Trauma Recovery." IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 2018. doi:10.1109/WACVW.2018.00016

Markdown

[Dhamija and Boult. "Learning Visual Engagement for Trauma Recovery." IEEE/CVF Winter Conference on Applications of Computer Vision Workshops, 2018.](https://mlanthology.org/wacvw/2018/dhamija2018wacvw-learning/) doi:10.1109/WACVW.2018.00016

BibTeX

@inproceedings{dhamija2018wacvw-learning,
  title     = {{Learning Visual Engagement for Trauma Recovery}},
  author    = {Dhamija, Svati and Boult, Terrance E.},
  booktitle = {IEEE/CVF Winter Conference on Applications of Computer Vision Workshops},
  year      = {2018},
  pages     = {84-93},
  doi       = {10.1109/WACVW.2018.00016},
  url       = {https://mlanthology.org/wacvw/2018/dhamija2018wacvw-learning/}
}