DORSal: Diffusion for Object-Centric Representations of Scenes $\textit{et Al.}$
Abstract
Recent progress in 3D scene understanding enables scalable learning of representations across large datasets of diverse scenes. As a consequence, generalization to unseen scenes and objects, rendering novel views from just a single or a handful of input images, and controllable scene generation that supports editing, is now possible. However, training jointly on a large number of scenes typically compromises rendering quality when compared to single-scene optimized models such as NeRFs. In this paper, we leverage recent progress in diffusion models to equip 3D scene representation learning models with the ability to render high-fidelity novel views, while retaining benefits such as object-level scene editing to a large degree. In particular, we propose DORSal, which adapts a video diffusion architecture for 3D scene generation conditioned on frozen object-centric slot-based representations of scenes. On both complex synthetic multi-object scenes and on the real-world large-scale Street View dataset, we show that DORSal enables scalable neural rendering of 3D scenes with object-level editing and improves upon existing approaches.
Cite
Text
Jabri et al. "DORSal: Diffusion for Object-Centric Representations of Scenes $\textit{et Al.}$." International Conference on Learning Representations, 2024.Markdown
[Jabri et al. "DORSal: Diffusion for Object-Centric Representations of Scenes $\textit{et Al.}$." International Conference on Learning Representations, 2024.](https://mlanthology.org/iclr/2024/jabri2024iclr-dorsal/)BibTeX
@inproceedings{jabri2024iclr-dorsal,
title = {{DORSal: Diffusion for Object-Centric Representations of Scenes $\textit{et Al.}$}},
author = {Jabri, Allan and van Steenkiste, Sjoerd and Hoogeboom, Emiel and Sajjadi, Mehdi S. M. and Kipf, Thomas},
booktitle = {International Conference on Learning Representations},
year = {2024},
url = {https://mlanthology.org/iclr/2024/jabri2024iclr-dorsal/}
}