Multi-Space Alignments Towards Universal LiDAR Segmentation
Abstract
A unified and versatile LiDAR segmentation model with strong robustness and generalizability is desirable for safe autonomous driving perception. This work presents M3Net a one-of-a-kind framework for fulfilling multi-task multi-dataset multi-modality LiDAR segmentation in a universal manner using just a single set of parameters. To better exploit data volume and diversity we first combine large-scale driving datasets acquired by different types of sensors from diverse scenes and then conduct alignments in three spaces namely data feature and label spaces during the training. As a result M3Net is capable of taming heterogeneous data for training state-of-the-art LiDAR segmentation models. Extensive experiments on twelve LiDAR segmentation datasets verify our effectiveness. Notably using a shared set of parameters M3Net achieves 75.1% 83.1% and 72.4% mIoU scores respectively on the official benchmarks of SemanticKITTI nuScenes and Waymo Open.
Cite
Text
Liu et al. "Multi-Space Alignments Towards Universal LiDAR Segmentation." Conference on Computer Vision and Pattern Recognition, 2024. doi:10.1109/CVPR52733.2024.01388Markdown
[Liu et al. "Multi-Space Alignments Towards Universal LiDAR Segmentation." Conference on Computer Vision and Pattern Recognition, 2024.](https://mlanthology.org/cvpr/2024/liu2024cvpr-multispace/) doi:10.1109/CVPR52733.2024.01388BibTeX
@inproceedings{liu2024cvpr-multispace,
title = {{Multi-Space Alignments Towards Universal LiDAR Segmentation}},
author = {Liu, Youquan and Kong, Lingdong and Wu, Xiaoyang and Chen, Runnan and Li, Xin and Pan, Liang and Liu, Ziwei and Ma, Yuexin},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2024},
pages = {14648-14661},
doi = {10.1109/CVPR52733.2024.01388},
url = {https://mlanthology.org/cvpr/2024/liu2024cvpr-multispace/}
}