NeRFusion: Fusing Radiance Fields for Large-Scale Scene Reconstruction

Abstract

While NeRF has shown great success for neural reconstruction and rendering, its limited MLP capacity and long per-scene optimization times make it challenging to model large-scale indoor scenes. In contrast, classical 3D reconstruction methods can handle large-scale scenes but do not produce realistic renderings. We propose NeRFusion, a method that combines the advantages of NeRF and TSDF-based fusion techniques to achieve efficient large-scale reconstruction and photo-realistic rendering. We process the input image sequence to predict per-frame local radiance fields via direct network inference. These are then fused using a novel recurrent neural network that incrementally reconstructs a global, sparse scene representation in real-time. This global volume can be further fine-tuned to boost rendering quality. We demonstrate that NeRFusion achieves state-of-the-art quality on both large-scale indoor and small-scale object scenes, with substantially faster reconstruction than NeRF and other recent methods.

Cite

Text

Zhang et al. "NeRFusion: Fusing Radiance Fields for Large-Scale Scene Reconstruction." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00537

Markdown

[Zhang et al. "NeRFusion: Fusing Radiance Fields for Large-Scale Scene Reconstruction." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/zhang2022cvpr-nerfusion/) doi:10.1109/CVPR52688.2022.00537

BibTeX

@inproceedings{zhang2022cvpr-nerfusion,
  title     = {{NeRFusion: Fusing Radiance Fields for Large-Scale Scene Reconstruction}},
  author    = {Zhang, Xiaoshuai and Bi, Sai and Sunkavalli, Kalyan and Su, Hao and Xu, Zexiang},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2022},
  pages     = {5449-5458},
  doi       = {10.1109/CVPR52688.2022.00537},
  url       = {https://mlanthology.org/cvpr/2022/zhang2022cvpr-nerfusion/}
}