HyperFields: Towards Zero-Shot Generation of NeRFs from Text

Abstract

We introduce HyperFields, a method for generating text-conditioned Neural Radiance Fields (NeRFs) with a single forward pass and (optionally) some fine-tuning. Key to our approach are: (i) a dynamic hypernetwork, which learns a smooth mapping from text token embeddings to the space of NeRFs; (ii) NeRF distillation training, which distills scenes encoded in individual NeRFs into one dynamic hypernetwork. These techniques enable a single network to fit over a hundred unique scenes. We further demonstrate that HyperFields learns a more general map between text and NeRFs, and consequently is capable of predicting novel in-distribution and out-of-distribution scenes — either zero-shot or with a few finetuning steps. Finetuning HyperFields benefits from accelerated convergence thanks to the learned general map, and is capable of synthesizing novel scenes 5 to 10 times faster than existing neural optimization-based methods. Our ablation experiments show that both the dynamic architecture and NeRF distillation are critical to the expressivity of HyperFields.

Cite

Text

Babu et al. "HyperFields: Towards Zero-Shot Generation of NeRFs from Text." International Conference on Machine Learning, 2024.

Markdown

[Babu et al. "HyperFields: Towards Zero-Shot Generation of NeRFs from Text." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/babu2024icml-hyperfields/)

BibTeX

@inproceedings{babu2024icml-hyperfields,
  title     = {{HyperFields: Towards Zero-Shot Generation of NeRFs from Text}},
  author    = {Babu, Sudarshan and Liu, Richard and Zhou, Avery and Maire, Michael and Shakhnarovich, Greg and Hanocka, Rana},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {2230-2247},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/babu2024icml-hyperfields/}
}