RICCARDO: Radar Hit Prediction and Convolution for Camera-Radar 3D Object Detection

Abstract

Radar hits reflect from points on both the boundary and internal to object outlines. This results in a complex distribution of radar hits that depends on factors including object category, size and orientation. Current radar-camera fusion methods implicitly account for this with a black-box neural network. In this paper, we explicitly utilize a radar hit distribution model to assist fusion. First, we build a model to predict radar hit distributions conditioned on object properties obtained from a monocular detector. Second, we use the predicted distribution as a kernel to match actual measured radar points in the neighborhood of the monocular detections, generating matching scores at nearby positions. Finally, a fusion stage combines context with the kernel detector to refine the matching scores. Our method achieves the state-of-the-art radar-camera detection performance on nuScenes. Our source code is available at https://github.com/longyunf/riccardo.

Cite

Text

Long et al. "RICCARDO: Radar Hit Prediction and Convolution for Camera-Radar 3D Object Detection." Conference on Computer Vision and Pattern Recognition, 2025. doi:10.1109/CVPR52734.2025.02075

Markdown

[Long et al. "RICCARDO: Radar Hit Prediction and Convolution for Camera-Radar 3D Object Detection." Conference on Computer Vision and Pattern Recognition, 2025.](https://mlanthology.org/cvpr/2025/long2025cvpr-riccardo/) doi:10.1109/CVPR52734.2025.02075

BibTeX

@inproceedings{long2025cvpr-riccardo,
  title     = {{RICCARDO: Radar Hit Prediction and Convolution for Camera-Radar 3D Object Detection}},
  author    = {Long, Yunfei and Kumar, Abhinav and Liu, Xiaoming and Morris, Daniel},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2025},
  pages     = {22276-22285},
  doi       = {10.1109/CVPR52734.2025.02075},
  url       = {https://mlanthology.org/cvpr/2025/long2025cvpr-riccardo/}
}