EmoCAM: Toward Understanding What Drives CNN-Based Emotion Recognition

Abstract

Convolutional Neural Networks are particularly suited for image analysis tasks, such as Image Classification, Object Recognition or Image Segmentation. Like all Artificial Neural Networks, however, they are "black box" models, and suffer from poor explainability. This work is concerned with the specific downstream task of Emotion Recognition from images, and proposes a framework that combines CAM-based techniques with Object Detection on a corpus level to better understand on which image cues a particular model relies to assign a specific emotion to an image. We demonstrate our framework using the EmoNet model and show that it mostly focuses on human characteristics, but also explore the pronounced effect of specific image modifications.

Cite

Text

Doulfoukar et al. "EmoCAM: Toward Understanding What Drives CNN-Based Emotion Recognition." NeurIPS 2024 Workshops: SciForDL, 2024.

Markdown

[Doulfoukar et al. "EmoCAM: Toward Understanding What Drives CNN-Based Emotion Recognition." NeurIPS 2024 Workshops: SciForDL, 2024.](https://mlanthology.org/neuripsw/2024/doulfoukar2024neuripsw-emocam/)

BibTeX

@inproceedings{doulfoukar2024neuripsw-emocam,
  title     = {{EmoCAM: Toward Understanding What Drives CNN-Based Emotion Recognition}},
  author    = {Doulfoukar, Youssef and Mertens, Laurent and Vennekens, Joost},
  booktitle = {NeurIPS 2024 Workshops: SciForDL},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/doulfoukar2024neuripsw-emocam/}
}