Light Field Neural Rendering
Abstract
Classical light field rendering for novel view synthesis can accurately reproduce view-dependent effects such as reflection, refraction, and translucency, but requires a dense view sampling of the scene. Methods based on geometric reconstruction need only sparse views, but cannot accurately model non-Lambertian effects. We introduce a model that combines the strengths and mitigates the limitations of these two directions. By operating on a four-dimensional representation of the light field, our model learns to represent view-dependent effects accurately. By enforcing geometric constraints during training and inference, the scene geometry is implicitly learned from a sparse set of views. Concretely, we introduce a two-stage transformer-based model that first aggregates features along epipolar lines, then aggregates features along reference views to produce the color of a target ray. Our model outperforms the state-of-the-art on multiple forward-facing and 360deg datasets, with larger margins on scenes with severe view-dependent variations.
Cite
Text
Suhail et al. "Light Field Neural Rendering." Conference on Computer Vision and Pattern Recognition, 2022. doi:10.1109/CVPR52688.2022.00809Markdown
[Suhail et al. "Light Field Neural Rendering." Conference on Computer Vision and Pattern Recognition, 2022.](https://mlanthology.org/cvpr/2022/suhail2022cvpr-light/) doi:10.1109/CVPR52688.2022.00809BibTeX
@inproceedings{suhail2022cvpr-light,
title = {{Light Field Neural Rendering}},
author = {Suhail, Mohammed and Esteves, Carlos and Sigal, Leonid and Makadia, Ameesh},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2022},
pages = {8269-8279},
doi = {10.1109/CVPR52688.2022.00809},
url = {https://mlanthology.org/cvpr/2022/suhail2022cvpr-light/}
}