Make-a-Shape: A Ten-Million-Scale 3D Shape Model
Abstract
The progression in large-scale 3D generative models has been impeded by significant resource requirements for training and challenges like inefficient representations. This paper introduces Make-A-Shape, a novel 3D generative model trained on a vast scale, using 10 million publicly-available shapes. We first innovate the wavelet-tree representation to encode high-resolution SDF shapes with minimal loss, leveraging our newly-proposed subband coefficient filtering scheme. We then design a subband coefficient packing scheme to facilitate diffusion-based generation and a subband adaptive training strategy for effective training on the large-scale dataset. Our generative framework is versatile, capable of conditioning on various input modalities such as images, point clouds, and voxels, enabling a variety of downstream applications, e.g., unconditional generation, completion, and conditional generation. Our approach clearly surpasses the existing baselines in delivering high-quality results and can efficiently generate shapes within two seconds for most conditions.
Cite
Text
Hui et al. "Make-a-Shape: A Ten-Million-Scale 3D Shape Model." International Conference on Machine Learning, 2024.Markdown
[Hui et al. "Make-a-Shape: A Ten-Million-Scale 3D Shape Model." International Conference on Machine Learning, 2024.](https://mlanthology.org/icml/2024/hui2024icml-makeashape/)BibTeX
@inproceedings{hui2024icml-makeashape,
title = {{Make-a-Shape: A Ten-Million-Scale 3D Shape Model}},
author = {Hui, Ka-Hei and Sanghi, Aditya and Rampini, Arianna and Rahimi Malekshan, Kamal and Liu, Zhengzhe and Shayani, Hooman and Fu, Chi-Wing},
booktitle = {International Conference on Machine Learning},
year = {2024},
pages = {20660-20681},
volume = {235},
url = {https://mlanthology.org/icml/2024/hui2024icml-makeashape/}
}