4D Panoptic LiDAR Segmentation
Abstract
Temporal semantic scene understanding is critical for self-driving cars or robots operating in dynamic environments. In this paper, we propose 4D panoptic LiDAR segmentation to assign a semantic class and a temporally-consistent instance ID to a sequence of 3D points. To this end, we present an approach and a novel evaluation metric. Our approach determines a semantic class for every point while modeling object instances as probability distributions in the 4D spatio-temporal domain. We process multiple point clouds in parallel and resolve point-to-instance associations, effectively alleviating the need for explicit temporal data association. Inspired by recent advances in benchmarking of multi-object tracking, we propose to adopt a new evaluation metric that separates the semantic and point-to-instance association aspects of the task. With this work, we aim at paving the road for future developments aiming at temporal LiDAR panoptic perception.
Cite
Text
Aygun et al. "4D Panoptic LiDAR Segmentation." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.00548Markdown
[Aygun et al. "4D Panoptic LiDAR Segmentation." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/aygun2021cvpr-4d/) doi:10.1109/CVPR46437.2021.00548BibTeX
@inproceedings{aygun2021cvpr-4d,
title = {{4D Panoptic LiDAR Segmentation}},
author = {Aygun, Mehmet and Osep, Aljosa and Weber, Mark and Maximov, Maxim and Stachniss, Cyrill and Behley, Jens and Leal-Taixe, Laura},
booktitle = {Conference on Computer Vision and Pattern Recognition},
year = {2021},
pages = {5527-5537},
doi = {10.1109/CVPR46437.2021.00548},
url = {https://mlanthology.org/cvpr/2021/aygun2021cvpr-4d/}
}