Multi-Level Neural Scene Graphs for Dynamic Urban Environments

Abstract

We estimate the radiance field of large-scale dynamic areas from multiple vehicle captures under varying environmental conditions. Previous works in this domain are either restricted to static environments do not scale to more than a single short video or struggle to separately represent dynamic object instances. To this end we present a novel decomposable radiance field approach for dynamic urban environments. We propose a multi-level neural scene graph representation that scales to thousands of images from dozens of sequences with hundreds of fast-moving objects. To enable efficient training and rendering of our representation we develop a fast composite ray sampling and rendering scheme. To test our approach in urban driving scenarios we introduce a new novel view synthesis benchmark. We show that our approach outperforms prior art by a significant margin on both established and our proposed benchmark while being faster in training and rendering.

Cite

Text

Fischer et al. "Multi-Level Neural Scene Graphs for Dynamic Urban Environments." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.01996

Markdown

[Fischer et al. "Multi-Level Neural Scene Graphs for Dynamic Urban Environments." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/fischer2024cvpr-multilevel/) doi:10.1109/CVPR52733.2024.01996

BibTeX

@inproceedings{fischer2024cvpr-multilevel,
  title     = {{Multi-Level Neural Scene Graphs for Dynamic Urban Environments}},
  author    = {Fischer, Tobias and Porzi, Lorenzo and Bulo, Samuel Rota and Pollefeys, Marc and Kontschieder, Peter},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2024},
  pages     = {21125-21135},
  doi       = {10.1109/CVPR52733.2024.01996},
  url       = {https://mlanthology.org/cvpr/2024/fischer2024cvpr-multilevel/}
}