ARF: Artistic Radiance Fields
Abstract
We present a method for transferring the artistic features of an arbitrary style image to a 3D scene. Previous methods that perform 3D stylization on point clouds or meshes are sensitive to geometric reconstruction errors for complex real-world scenes. Instead, we propose to stylize the more robust radiance field representation. We find that the commonly used Gram matrix-based loss tends to produce blurry results lacking in faithful style detail. We instead utilize a nearest neighbor-based loss that is highly effective at capturing style details while maintaining multi-view consistency. We also propose a novel deferred back-propagation method to enable optimization of memory-intensive radiance fields using style losses defined on full-resolution rendered images. Our evaluation demonstrates that, compared to baselines, our method transfers artistic appearance in a way that more closely resembles the style image. Please see our project webpage for video results and an open-source implementation: https://www.cs.cornell.edu/projects/arf/.
Cite
Text
Zhang et al. "ARF: Artistic Radiance Fields." Proceedings of the European Conference on Computer Vision (ECCV), 2022. doi:10.1007/978-3-031-19821-2_41Markdown
[Zhang et al. "ARF: Artistic Radiance Fields." Proceedings of the European Conference on Computer Vision (ECCV), 2022.](https://mlanthology.org/eccv/2022/zhang2022eccv-arf/) doi:10.1007/978-3-031-19821-2_41BibTeX
@inproceedings{zhang2022eccv-arf,
title = {{ARF: Artistic Radiance Fields}},
author = {Zhang, Kai and Kolkin, Nick and Bi, Sai and Luan, Fujun and Xu, Zexiang and Shechtman, Eli and Snavely, Noah},
booktitle = {Proceedings of the European Conference on Computer Vision (ECCV)},
year = {2022},
doi = {10.1007/978-3-031-19821-2_41},
url = {https://mlanthology.org/eccv/2022/zhang2022eccv-arf/}
}