GEOPARD: Geometric Pretraining for Articulation Prediction in 3D Shapes
Abstract
We present GEOPARD, a transformer-based architecture for predicting articulation from a single static snapshot of a 3D shape. The key idea of our method is a pretraining strategy that allows our transformer to learn plausible candidate articulations for 3D shapes based on a geometric-driven search without manual articulation annotation. The search automatically discovers physically valid part motions that do not cause detachments or collisions with other shape parts. Our experiments indicate that this geometric pretraining strategy, along with carefully designed choices in our transformer architecture, yields state-of-the-art results in articulation inference in the popular shape Part-Mobility dataset.
Cite
Text
Goyal et al. "GEOPARD: Geometric Pretraining for Articulation Prediction in 3D Shapes." International Conference on Computer Vision, 2025.Markdown
[Goyal et al. "GEOPARD: Geometric Pretraining for Articulation Prediction in 3D Shapes." International Conference on Computer Vision, 2025.](https://mlanthology.org/iccv/2025/goyal2025iccv-geopard/)BibTeX
@inproceedings{goyal2025iccv-geopard,
title = {{GEOPARD: Geometric Pretraining for Articulation Prediction in 3D Shapes}},
author = {Goyal, Pradyumn and Petrov, Dmitry and Andrews, Sheldon and Ben-Shabat, Yizhak and Liu, Hsueh-Ti Derek and Kalogerakis, Evangelos},
booktitle = {International Conference on Computer Vision},
year = {2025},
pages = {9332-9341},
url = {https://mlanthology.org/iccv/2025/goyal2025iccv-geopard/}
}