SoundBrush: Sound as a Brush for Visual Scene Editing
Abstract
We propose SoundBrush, a model that uses sound as a brush to edit and manipulate visual scenes. We extend the generative capabilities of the Latent Diffusion Model (LDM) to incorporate audio information for editing visual scenes. Inspired by existing image-editing works, we frame this task as a supervised learning problem and leverage various off-the-shelf models to construct a sound-paired visual scene editing dataset for training. This richly generated dataset enables SoundBrush to learn to map audio features into the textual space of the LDM, allowing for visual scene editing guided by diverse in-the-wild sound. Unlike existing methods, SoundBrush can accurately manipulate the overall scenery or even insert sounding objects to best match the input sound semantics while preserving the original content. Furthermore, by integrating with novel view synthesis techniques, our framework can be extended to edit 3D scenes, facilitating sound-driven 3D scene manipulation.
Cite
Text
Sung-Bin et al. "SoundBrush: Sound as a Brush for Visual Scene Editing." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I7.32770Markdown
[Sung-Bin et al. "SoundBrush: Sound as a Brush for Visual Scene Editing." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/sungbin2025aaai-soundbrush/) doi:10.1609/AAAI.V39I7.32770BibTeX
@inproceedings{sungbin2025aaai-soundbrush,
title = {{SoundBrush: Sound as a Brush for Visual Scene Editing}},
author = {Sung-Bin, Kim and Jun-Seong, Kim and Ko, Junseok and Kim, Yewon and Oh, Tae-Hyun},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2025},
pages = {7167-7175},
doi = {10.1609/AAAI.V39I7.32770},
url = {https://mlanthology.org/aaai/2025/sungbin2025aaai-soundbrush/}
}