Visual Instruction Inversion: Image Editing via Image Prompting
Abstract
Text-conditioned image editing has emerged as a powerful tool for editing images.However, in many situations, language can be ambiguous and ineffective in describing specific image edits.When faced with such challenges, visual prompts can be a more informative and intuitive way to convey ideas.We present a method for image editing via visual prompting.Given pairs of example that represent the "before" and "after" images of an edit, our goal is to learn a text-based editing direction that can be used to perform the same edit on new images.We leverage the rich, pretrained editing capabilities of text-to-image diffusion models by inverting visual prompts into editing instructions.Our results show that with just one example pair, we can achieve competitive results compared to state-of-the-art text-conditioned image editing frameworks.
Cite
Text
Nguyen et al. "Visual Instruction Inversion: Image Editing via Image Prompting." Neural Information Processing Systems, 2023.Markdown
[Nguyen et al. "Visual Instruction Inversion: Image Editing via Image Prompting." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/nguyen2023neurips-visual/)BibTeX
@inproceedings{nguyen2023neurips-visual,
title = {{Visual Instruction Inversion: Image Editing via Image Prompting}},
author = {Nguyen, Thao and Li, Yuheng and Ojha, Utkarsh and Lee, Yong Jae},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/nguyen2023neurips-visual/}
}