Audiovisual Conflict Detection in Political Debates
Abstract
In this paper, the problem of conflict detection in audiovisual recordings of political debates is investigated. In contrast to the current state of the art in social signal processing, where only the audio modality is employed for analysing the human non-verbal behavior, we propose to use additionally visual features capturing certain facial behavioral cues such as head nodding, fidgeting, and frowning which are related to conflicts. To this end, a dataset with video excerpts from televised political debates, where conflicts naturally arise, is introduced. The prediction of conflict level (i.e., conflict/nonconflict) is performed by applying the linear support vector machine and the collaborative representation-based classifier onto audio, visual, and audiovisual features. The experimental results demonstrate that the fusion of audio and visual features, outperform the accuracy in conflict detection, obtained by features that resort to a single modality (i.e., either audio or video).
Cite
Text
Panagakis et al. "Audiovisual Conflict Detection in Political Debates." European Conference on Computer Vision Workshops, 2014. doi:10.1007/978-3-319-16178-5_21Markdown
[Panagakis et al. "Audiovisual Conflict Detection in Political Debates." European Conference on Computer Vision Workshops, 2014.](https://mlanthology.org/eccvw/2014/panagakis2014eccvw-audiovisual/) doi:10.1007/978-3-319-16178-5_21BibTeX
@inproceedings{panagakis2014eccvw-audiovisual,
title = {{Audiovisual Conflict Detection in Political Debates}},
author = {Panagakis, Yannis and Zafeiriou, Stefanos and Pantic, Maja},
booktitle = {European Conference on Computer Vision Workshops},
year = {2014},
pages = {306-314},
doi = {10.1007/978-3-319-16178-5_21},
url = {https://mlanthology.org/eccvw/2014/panagakis2014eccvw-audiovisual/}
}