Facial Affect Estimation in the Wild Using Deep Residual and Convolutional Networks

Abstract

Automated affective computing in the wild is a challenging task in the field of computer vision. This paper presents three neural network-based methods proposed for the task of facial affect estimation submitted to the First Affect-in-the-Wild challenge. These methods are based on Inception-ResNet modules redesigned specifically for the task of facial affect estimation. These methods are: Shallow Inception-ResNet, Deep Inception-ResNet, and Inception-ResNet with LSTMs. These networks extract facial features in different scales and simultaneously estimate both the valence and arousal in each frame. Root Mean Square Error (RMSE) rates of 0.4 and 0.3 are achieved for the valence and arousal respectively with corresponding Concordance Correlation Coefficient (CCC) rates of 0.04 and 0.29 using Deep Inception-ResNet method.

Cite

Text

Hassani and Mahoor. "Facial Affect Estimation in the Wild Using Deep Residual and Convolutional Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017. doi:10.1109/CVPRW.2017.245

Markdown

[Hassani and Mahoor. "Facial Affect Estimation in the Wild Using Deep Residual and Convolutional Networks." IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2017.](https://mlanthology.org/cvprw/2017/hassani2017cvprw-facial/) doi:10.1109/CVPRW.2017.245

BibTeX

@inproceedings{hassani2017cvprw-facial,
  title     = {{Facial Affect Estimation in the Wild Using Deep Residual and Convolutional Networks}},
  author    = {Hassani, Behzad and Mahoor, Mohammad H.},
  booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops},
  year      = {2017},
  pages     = {1955-1962},
  doi       = {10.1109/CVPRW.2017.245},
  url       = {https://mlanthology.org/cvprw/2017/hassani2017cvprw-facial/}
}