SoundBrush: Sound as a Brush for Visual Scene Editing

Abstract

We propose SoundBrush, a model that uses sound as a brush to edit and manipulate visual scenes. We extend the generative capabilities of the Latent Diffusion Model (LDM) to incorporate audio information for editing visual scenes. Inspired by existing image-editing works, we frame this task as a supervised learning problem and leverage various off-the-shelf models to construct a sound-paired visual scene editing dataset for training. This richly generated dataset enables SoundBrush to learn to map audio features into the textual space of the LDM, allowing for visual scene editing guided by diverse in-the-wild sound. Unlike existing methods, SoundBrush can accurately manipulate the overall scenery or even insert sounding objects to best match the input sound semantics while preserving the original content. Furthermore, by integrating with novel view synthesis techniques, our framework can be extended to edit 3D scenes, facilitating sound-driven 3D scene manipulation.

Cite

Text

Sung-Bin et al. "SoundBrush: Sound as a Brush for Visual Scene Editing." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I7.32770

Markdown

[Sung-Bin et al. "SoundBrush: Sound as a Brush for Visual Scene Editing." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/sungbin2025aaai-soundbrush/) doi:10.1609/AAAI.V39I7.32770

BibTeX

@inproceedings{sungbin2025aaai-soundbrush,
  title     = {{SoundBrush: Sound as a Brush for Visual Scene Editing}},
  author    = {Sung-Bin, Kim and Jun-Seong, Kim and Ko, Junseok and Kim, Yewon and Oh, Tae-Hyun},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {7167-7175},
  doi       = {10.1609/AAAI.V39I7.32770},
  url       = {https://mlanthology.org/aaai/2025/sungbin2025aaai-soundbrush/}
}