PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer

Abstract

3D surface reconstruction from unorganized point clouds is a fundamental task in visual computing with numerous applications in areas such as robotics virtual reality augmented reality and animation. To date many deep learning-based surface reconstruction methods have been proposed demonstrating great performance on many benchmark datasets. Among these neural implicit field learning-based methods have gained popularity for their capability of representing complex structures in a continuous implicit distance field. Existing neural implicit field learning methods either utilize voxelized point cloud then feed them to a deep network or directly take points as input. In this paper we propose an implicit surface reconstruction framework based on point voxel geometric-aware transformer PVT to seamlessly integrate point-based convolution with voxel-based convolution using bidirectional transformers. Experiments show that the proposed PVT framework can better encode local geometry details and provide a significant performance boost over existing state-of-the-art methods.

Cite

Text

Fan et al. "PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer." Winter Conference on Applications of Computer Vision, 2025.

Markdown

[Fan et al. "PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer." Winter Conference on Applications of Computer Vision, 2025.](https://mlanthology.org/wacv/2025/fan2025wacv-pvt/)

BibTeX

@inproceedings{fan2025wacv-pvt,
  title     = {{PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer}},
  author    = {Fan, Chuanmao and Zhao, Chenxi and Duan, Ye},
  booktitle = {Winter Conference on Applications of Computer Vision},
  year      = {2025},
  pages     = {3013-3023},
  url       = {https://mlanthology.org/wacv/2025/fan2025wacv-pvt/}
}