MLADDC: Multi-Lingual Audio Deepfake Detection Corpus

Abstract

This study develop Multi-Lingual Audio Deepfake Detection Corpus (MLADDC) to boost the ADD research. Existing datasets suffer from several limitations, in particular, they are limited to one or two languages. Proposed dataset contains 20 languages, which have been released in 4 Tracks (6 - Indian languages, 14 - International languages, 20 languages half-truth data, and combined data). Moreover, the proposed dataset has 400K files (1,125+ hours) of data, which makes it one of the largest datasets. Deepfakes in MLADDC have been produced using advanced DL methods, such as HiFiGAN and BigVGAN. Another novelty lies in its sub-dataset, that has partial deepfakes (Half-Truth). We compared our dataset with various existing datasets, using cross-database method. For comparison, we also proposed baseline accuracy of 68.44 %, and EER of 40.9 % with MFCC features and CNN classifier (14 languages track only) indicating technological challenges associated with ADD task on proposed dataset.

Cite

Text

Shah et al. "MLADDC: Multi-Lingual Audio Deepfake Detection Corpus." NeurIPS 2024 Workshops: Audio_Imagination, 2024.

Markdown

[Shah et al. "MLADDC: Multi-Lingual Audio Deepfake Detection Corpus." NeurIPS 2024 Workshops: Audio_Imagination, 2024.](https://mlanthology.org/neuripsw/2024/shah2024neuripsw-mladdc/)

BibTeX

@inproceedings{shah2024neuripsw-mladdc,
  title     = {{MLADDC: Multi-Lingual Audio Deepfake Detection Corpus}},
  author    = {Shah, Arth Juhul and Purohit, Ravindrakumar M. and Vaghera, Dharmendra H. and Patil, Hemant},
  booktitle = {NeurIPS 2024 Workshops: Audio_Imagination},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/shah2024neuripsw-mladdc/}
}