TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/CRN: Camera Radar Net for Accurate, Robust, Efficient 3D P...

CRN: Camera Radar Net for Accurate, Robust, Efficient 3D Perception

Youngseok Kim, Juyeb Shin, Sanmin Kim, In-Jae Lee, Jun Won Choi, Dongsuk Kum

2023-04-03ICCV 2023 13D Object TrackingAutonomous Driving3D Multi-Object Trackingobject-detection3D Object DetectionObject Detection
PaperPDFCode(official)

Abstract

Autonomous driving requires an accurate and fast 3D perception system that includes 3D object detection, tracking, and segmentation. Although recent low-cost camera-based approaches have shown promising results, they are susceptible to poor illumination or bad weather conditions and have a large localization error. Hence, fusing camera with low-cost radar, which provides precise long-range measurement and operates reliably in all environments, is promising but has not yet been thoroughly investigated. In this paper, we propose Camera Radar Net (CRN), a novel camera-radar fusion framework that generates a semantically rich and spatially accurate bird's-eye-view (BEV) feature map for various tasks. To overcome the lack of spatial information in an image, we transform perspective view image features to BEV with the help of sparse but accurate radar points. We further aggregate image and radar feature maps in BEV using multi-modal deformable attention designed to tackle the spatial misalignment between inputs. CRN with real-time setting operates at 20 FPS while achieving comparable performance to LiDAR detectors on nuScenes, and even outperforms at a far distance on 100m setting. Moreover, CRN with offline setting yields 62.4% NDS, 57.5% mAP on nuScenes test set and ranks first among all camera and camera-radar 3D object detectors.

Results

TaskDatasetMetricValueModel
Multi-Object Trackingnuscenes Camera-RadarAMOTA0.569CRN
Multi-Object TrackingnuScenesAMOTA0.569CRN
Object Trackingnuscenes Camera-RadarAMOTA0.569CRN
Object TrackingnuScenesAMOTA0.569CRN
Object DetectionnuScenesNDS0.624CRN
Object DetectionnuScenesmAP0.575CRN
Object Detectionnuscenes Camera-RadarNDS62.4CRN
3DnuScenesNDS0.624CRN
3DnuScenesmAP0.575CRN
3Dnuscenes Camera-RadarNDS62.4CRN
3D Object DetectionnuScenesNDS0.624CRN
3D Object DetectionnuScenesmAP0.575CRN
3D Object Detectionnuscenes Camera-RadarNDS62.4CRN
3D Multi-Object Trackingnuscenes Camera-RadarAMOTA0.569CRN
3D Multi-Object TrackingnuScenesAMOTA0.569CRN
2D ClassificationnuScenesNDS0.624CRN
2D ClassificationnuScenesmAP0.575CRN
2D Classificationnuscenes Camera-RadarNDS62.4CRN
2D Object DetectionnuScenesNDS0.624CRN
2D Object DetectionnuScenesmAP0.575CRN
2D Object Detectionnuscenes Camera-RadarNDS62.4CRN
16knuScenesNDS0.624CRN
16knuScenesmAP0.575CRN
16knuscenes Camera-RadarNDS62.4CRN

Related Papers

GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17