Make-a-Shape: A Ten-Million-Scale 3D Shape Model

Abstract

The progression in large-scale 3D generative models has been impeded by significant resource requirements for training and challenges like inefficient representations. This paper introduces Make-A-Shape, a novel 3D generative model trained on a vast scale, using 10 million publicly-available shapes. We first innovate the wavelet-tree representation to encode high-resolution SDF shapes with minimal loss, leveraging our newly-proposed subband coefficient filtering scheme. We then design a subband coefficient packing scheme to facilitate diffusion-based generation and a subband adaptive training strategy for effective training on the large-scale dataset. Our generative framework is versatile, capable of conditioning on various input modalities such as images, point clouds, and voxels, enabling a variety of downstream applications, e.g., unconditional generation, completion, and conditional generation. Our approach clearly surpasses the existing baselines in delivering high-quality results and can efficiently generate shapes within two seconds for most conditions.

Cite

Text

Hui et al. "Make-a-Shape: A Ten-Million-Scale 3D Shape Model." International Conference on Machine Learning, 2024.

Markdown

[Hui et al. "Make-a-Shape: A Ten-Million-Scale 3D Shape Model." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/hui2024icml-makeashape/)

BibTeX

@inproceedings{hui2024icml-makeashape,
  title     = {{Make-a-Shape: A Ten-Million-Scale 3D Shape Model}},
  author    = {Hui, Ka-Hei and Sanghi, Aditya and Rampini, Arianna and Rahimi Malekshan, Kamal and Liu, Zhengzhe and Shayani, Hooman and Fu, Chi-Wing},
  booktitle = {International Conference on Machine Learning},
  year      = {2024},
  pages     = {20660-20681},
  volume    = {235},
  url       = {https://mlanthology.org/icml/2024/hui2024icml-makeashape/}
}