Diverse and Stable 2D Diffusion Guided Text to 3D Generation with Noise Recalibration

Abstract

In recent years, following the success of text guided image generation, text guided 3D generation has gained increasing attention among researchers. Dreamfusion is a notable approach that enhances generation quality by utilizing 2D text guided diffusion models and introducing SDS loss, a technique for distilling 2D diffusion model information to train 3D models. However, the SDS loss has two major limitations that hinder its effectiveness. Firstly, when given a text prompt, the SDS loss struggles to produce diverse content. Secondly, during training, SDS loss may cause the generated content to overfit and collapse, limiting the model's ability to learn intricate texture details. To overcome these challenges, we propose a novel approach called Noise Recalibration algorithm. By incorporating this technique, we can generate 3D content with significantly greater diversity and stunning details. Our approach offers a promising solution to the limitations of SDS loss.

Cite

Text

Yang et al. "Diverse and Stable 2D Diffusion Guided Text to 3D Generation with Noise Recalibration." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I7.28476

Markdown

[Yang et al. "Diverse and Stable 2D Diffusion Guided Text to 3D Generation with Noise Recalibration." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/yang2024aaai-diverse/) doi:10.1609/AAAI.V38I7.28476

BibTeX

@inproceedings{yang2024aaai-diverse,
  title     = {{Diverse and Stable 2D Diffusion Guided Text to 3D Generation with Noise Recalibration}},
  author    = {Yang, Xiaofeng and Liu, Fayao and Xu, Yi and Su, Hanjing and Wu, Qingyao and Lin, Guosheng},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {6549-6557},
  doi       = {10.1609/AAAI.V38I7.28476},
  url       = {https://mlanthology.org/aaai/2024/yang2024aaai-diverse/}
}