DE-Net: Dynamic Text-Guided Image Editing Adversarial Networks
Abstract
Text-guided image editing models have shown remarkable results. However, there remain two problems. First, they employ fixed manipulation modules for various editing requirements (e.g., color changing, texture changing, content adding and removing), which results in over-editing or insufficient editing. Second, they do not clearly distinguish between text-required and text-irrelevant parts, which leads to inaccurate editing. To solve these limitations, we propose: (i) a Dynamic Editing Block (DEBlock) that composes different editing modules dynamically for various editing requirements. (ii) a Composition Predictor (Comp-Pred), which predicts the composition weights for DEBlock according to the inference on target texts and source images. (iii) a Dynamic text-adaptive Convolution Block (DCBlock) that queries source image features to distinguish text-required parts and text-irrelevant parts. Extensive experiments demonstrate that our DE-Net achieves excellent performance and manipulates source images more correctly and accurately.
Cite
Text
Tao et al. "DE-Net: Dynamic Text-Guided Image Editing Adversarial Networks." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I8.26189Markdown
[Tao et al. "DE-Net: Dynamic Text-Guided Image Editing Adversarial Networks." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/tao2023aaai-de/) doi:10.1609/AAAI.V37I8.26189BibTeX
@inproceedings{tao2023aaai-de,
title = {{DE-Net: Dynamic Text-Guided Image Editing Adversarial Networks}},
author = {Tao, Ming and Bao, Bing-Kun and Tang, Hao and Wu, Fei and Wei, Longhui and Tian, Qi},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2023},
pages = {9971-9979},
doi = {10.1609/AAAI.V37I8.26189},
url = {https://mlanthology.org/aaai/2023/tao2023aaai-de/}
}