AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea

Abstract

Instruction-based image editing aims to modify specific image elements with natural language instructions. However, current models in this domain often struggle to execute complex user instructions accurately, as they are trained on low-quality data with limited editing types. We present AnyEdit, a comprehensive multi-modal instruction editing dataset, comprising 2.5 million high-quality editing pairs spanning over 20 editing types and five domains. We ensure the diversity and quality of the AnyEdit collection through three aspects: initial data diversity, adaptive editing process, and automated selection of editing results. Using the dataset, we further train a novel AnyEdit Stable Diffusion with task-aware routing and learnable task embedding for unified image editing. Comprehensive experiments on three benchmark datasets show that AnyEdit consistently boosts the performance of diffusion-based editing models. This presents prospects for developing instruction-driven image editing models that support human creativity.

Cite

Text

Yu et al. "AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.02433

Markdown

[Yu et al. "AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/yu2025cvpr-anyedit/) doi:10.1109/CVPR52734.2025.02433

BibTeX

@inproceedings{yu2025cvpr-anyedit,
  title     = {{AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea}},
  author    = {Yu, Qifan and Chow, Wei and Yue, Zhongqi and Pan, Kaihang and Wu, Yang and Wan, Xiaoyang and Li, Juncheng and Tang, Siliang and Zhang, Hanwang and Zhuang, Yueting},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {26125-26135},
  doi       = {10.1109/CVPR52734.2025.02433},
  url       = {https://mlanthology.org/cvpr/2025/yu2025cvpr-anyedit/}
}