Training-Free Image Manipulation Localization Using Diffusion Models
Abstract
Image manipulation localization (IML) is a critical technique in media forensics, focusing on identifying tampered regions within manipulated images. Most existing IML methods require extensive training on labeled datasets with both image-level and pixel-level annotations. These methods often struggle with new manipulation types and exhibit low generalizability. In this work, we propose a training-free IML approach using diffusion models. Our method adaptively selects an appropriate number of diffusion timesteps for each input image in the forward process and performs both conditional and unconditional reconstructions in the backward process without relying on external conditions. By comparing these reconstructions, we generate a localization map highlighting regions of manipulation based on inconsistencies. Extensive experiments were conducted using sixteen state-of-the-art (SoTA) methods across six IML datasets. The results demonstrate that our training-free method outperforms SoTA unsupervised and weakly-supervised techniques. Furthermore, our method competes effectively against fully-supervised methods on novel (unseen) manipulation types.
Cite
Text
Zhang et al. "Training-Free Image Manipulation Localization Using Diffusion Models." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I10.33126Markdown
[Zhang et al. "Training-Free Image Manipulation Localization Using Diffusion Models." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/zhang2025aaai-training/) doi:10.1609/AAAI.V39I10.33126BibTeX
@inproceedings{zhang2025aaai-training,
title = {{Training-Free Image Manipulation Localization Using Diffusion Models}},
author = {Zhang, Zhenfei and Chang, Ming-Ching and Li, Xin},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {10376-10384},
doi = {10.1609/AAAI.V39I10.33126},
url = {https://mlanthology.org/aaai/2025/zhang2025aaai-training/}
}