OctFormer: Efficient Octree-Based Transformer for Point Cloud Compression with Local Enhancement

Abstract

Point cloud compression with a higher compression ratio and tiny loss is essential for efficient data transportation. However, previous methods that depend on 3D convolution or frequent multi-head self-attention operations bring huge computations. To address this problem, we propose an octree-based Transformer compression method called OctFormer, which does not rely on the occupancy information of sibling nodes. Our method uses non-overlapped context windows to construct octree node sequences and share the result of a multi-head self-attention operation among a sequence of nodes. Besides, we introduce a locally-enhance module for exploiting the sibling features and a positional encoding generator for enhancing the translation invariance of the octree node sequence. Compared to the previous state-of-the-art works, our method obtains up to 17% Bpp savings compared to the voxel-context-based baseline and saves an overall 99% coding time compared to the attention-based baseline.

Cite

Text

Cui et al. "OctFormer: Efficient Octree-Based Transformer for Point Cloud Compression with Local Enhancement." AAAI Conference on Artificial Intelligence, 2023. doi:10.1609/AAAI.V37I1.25121

Markdown

[Cui et al. "OctFormer: Efficient Octree-Based Transformer for Point Cloud Compression with Local Enhancement." AAAI Conference on Artificial Intelligence, 2023.](https://mlanthology.org/aaai/2023/cui2023aaai-octformer/) doi:10.1609/AAAI.V37I1.25121

BibTeX

@inproceedings{cui2023aaai-octformer,
  title     = {{OctFormer: Efficient Octree-Based Transformer for Point Cloud Compression with Local Enhancement}},
  author    = {Cui, Mingyue and Long, Junhua and Feng, Mingjian and Li, Boyang and Huang, Kai},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2023},
  pages     = {470-478},
  doi       = {10.1609/AAAI.V37I1.25121},
  url       = {https://mlanthology.org/aaai/2023/cui2023aaai-octformer/}
}