Hidden Gems: 4D Radar Scene Flow Learning Using Cross-Modal Supervision

Abstract

This work proposes a novel approach to 4D radar-based scene flow estimation via cross-modal learning. Our approach is motivated by the co-located sensing redundancy in modern autonomous vehicles. Such redundancy implicitly provides various forms of supervision cues to the radar scene flow estimation. Specifically, we introduce a multi-task model architecture for the identified cross-modal learning problem and propose loss functions to opportunistically engage scene flow estimation using multiple cross-modal constraints for effective model training. Extensive experiments show the state-of-the-art performance of our method and demonstrate the effectiveness of cross-modal supervised learning to infer more accurate 4D radar scene flow. We also show its usefulness to two subtasks - motion segmentation and ego-motion estimation. Our source code will be available on https://github.com/Toytiny/CMFlow.

Cite

Text

Ding et al. "Hidden Gems: 4D Radar Scene Flow Learning Using Cross-Modal Supervision." Conference on Computer Vision and Pattern Recognition, 2023. doi:10.1109/CVPR52729.2023.00901

Markdown

[Ding et al. "Hidden Gems: 4D Radar Scene Flow Learning Using Cross-Modal Supervision." Conference on Computer Vision and Pattern Recognition, 2023.](https://mlanthology.org/cvpr/2023/ding2023cvpr-hidden/) doi:10.1109/CVPR52729.2023.00901

BibTeX

@inproceedings{ding2023cvpr-hidden,
  title     = {{Hidden Gems: 4D Radar Scene Flow Learning Using Cross-Modal Supervision}},
  author    = {Ding, Fangqiang and Palffy, Andras and Gavrila, Dariu M. and Lu, Chris Xiaoxuan},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2023},
  pages     = {9340-9349},
  doi       = {10.1109/CVPR52729.2023.00901},
  url       = {https://mlanthology.org/cvpr/2023/ding2023cvpr-hidden/}
}