Neural Residual Radiance Fields for Streamably Free-Viewpoint Videos
Abstract
The success of the Neural Radiance Fields (NeRFs) for modeling and free-view rendering static objects has inspired numerous attempts on dynamic scenes. Current techniques that utilize neural rendering for facilitating free-view videos (FVVs) are restricted to either offline rendering or are capable of processing only brief sequences with minimal motion. In this paper, we present a novel technique, Residual Radiance Field or ReRF, as a highly compact neural representation to achieve real-time FVV rendering on long-duration dynamic scenes. ReRF explicitly models the residual information between adjacent timestamps in the spatial-temporal feature space, with a global coordinate-based tiny MLP as the feature decoder. Specifically, ReRF employs a compact motion grid along with a residual feature grid to exploit inter-frame feature similarities. We show such a strategy can handle large motions without sacrificing quality. We further present a sequential training scheme to maintain the smoothness and the sparsity of the motion/residual grids. Based on ReRF, we design a special FVV codec that achieves three orders of magnitudes compression rate and provides a companion ReRF player to support online streaming of long-duration FVVs of dynamic scenes. Extensive experiments demonstrate the effectiveness of ReRF for compactly representing dynamic radiance fields, enabling an unprecedented free-viewpoint viewing experience in speed and quality.
Cite
Text
Wang et al. "Neural Residual Radiance Fields for Streamably Free-Viewpoint Videos." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.00016Markdown
[Wang et al. "Neural Residual Radiance Fields for Streamably Free-Viewpoint Videos." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/wang2023cvpr-neural/) doi:10.1109/CVPR52729.2023.00016BibTeX
@inproceedings{wang2023cvpr-neural,
title = {{Neural Residual Radiance Fields for Streamably Free-Viewpoint Videos}},
author = {Wang, Liao and Hu, Qiang and He, Qihan and Wang, Ziyu and Yu, Jingyi and Tuytelaars, Tinne and Xu, Lan and Wu, Minye},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2023},
pages = {76-87},
doi = {10.1109/CVPR52729.2023.00016},
url = {https://mlanthology.org/cvpr/2023/wang2023cvpr-neural/}
}