Lognormal Mutations and Their Use in Detecting Surreptitious Fake Images

Abstract

In many cases, adversarial attacks against fake detectors employ algorithms specifically crafted for automatic image classifiers. These algorithms perform well, thanks to an excellent ad hoc distribution of initial attacks. However, these attacks are easily detected due to their specific initial distribution. Consequently, we explore alternative black-box attacks inspired by generic black-box optimization tools, particularly focusing on the \lognormal{} algorithm that we successfully extend to attack fake detectors. Moreover, we demonstrate that this attack evades detection by neural networks trained to flag classical adversarial examples. Therefore, we train more general models capable of identifying a broader spectrum of attacks, including classical black-box attacks designed for images, black-box attacks driven by classical optimization, and no-box attacks. By integrating these attack detection capabilities with fake detectors, we develop more robust and effective fake detection systems.

Cite

Text

Teytaud et al. "Lognormal Mutations and Their Use in Detecting Surreptitious Fake Images." Transactions on Machine Learning Research, 2025.

Markdown

[Teytaud et al. "Lognormal Mutations and Their Use in Detecting Surreptitious Fake Images." Transactions on Machine Learning Research, 2025.](https://mlanthology.org/tmlr/2025/teytaud2025tmlr-lognormal/)

BibTeX

@article{teytaud2025tmlr-lognormal,
  title     = {{Lognormal Mutations and Their Use in Detecting Surreptitious Fake Images}},
  author    = {Teytaud, Olivier and Zameshina, Mariia and Sander, Tom and Fernandez, Pierre and Ye, Furong and Najman, Laurent and Bäck, Thomas and Labiad, Ismail},
  journal   = {Transactions on Machine Learning Research},
  year      = {2025},
  url       = {https://mlanthology.org/tmlr/2025/teytaud2025tmlr-lognormal/}
}