PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer
Abstract
3D surface reconstruction from unorganized point clouds is a fundamental task in visual computing with numerous applications in areas such as robotics virtual reality augmented reality and animation. To date many deep learning-based surface reconstruction methods have been proposed demonstrating great performance on many benchmark datasets. Among these neural implicit field learning-based methods have gained popularity for their capability of representing complex structures in a continuous implicit distance field. Existing neural implicit field learning methods either utilize voxelized point cloud then feed them to a deep network or directly take points as input. In this paper we propose an implicit surface reconstruction framework based on point voxel geometric-aware transformer PVT to seamlessly integrate point-based convolution with voxel-based convolution using bidirectional transformers. Experiments show that the proposed PVT framework can better encode local geometry details and provide a significant performance boost over existing state-of-the-art methods.
Cite
Text
Fan et al. "PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer." Winter Conference on Applications of Computer Vision, 2025.Markdown
[Fan et al. "PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer." Winter Conference on Applications of Computer Vision, 2025.](https://mlanthology.org/wacv/2025/fan2025wacv-pvt/)BibTeX
@inproceedings{fan2025wacv-pvt,
title = {{PVT: An Implicit Surface Reconstruction Framework via Point Voxel Geometric-Aware Transformer}},
author = {Fan, Chuanmao and Zhao, Chenxi and Duan, Ye},
booktitle = {Winter Conference on Applications of Computer Vision},
year = {2025},
pages = {3013-3023},
url = {https://mlanthology.org/wacv/2025/fan2025wacv-pvt/}
}