TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Unleashing HyDRa: Hybrid Fusion, Depth Consistency and Rad...

Unleashing HyDRa: Hybrid Fusion, Depth Consistency and Radar for Unified 3D Perception

Philipp Wolters, Johannes Gilg, Torben Teepe, Fabian Herzog, Anouar Laouichi, Martin Hofmann, Gerhard Rigoll

2024-03-12Prediction Of Occupancy Grid Maps3D Object Detection (RoI)3D Semantic Occupancy PredictionDepth PredictionAutonomous Driving3D Multi-Object Tracking3D Object Detection
PaperPDFCode(official)

Abstract

Low-cost, vision-centric 3D perception systems for autonomous driving have made significant progress in recent years, narrowing the gap to expensive LiDAR-based methods. The primary challenge in becoming a fully reliable alternative lies in robust depth prediction capabilities, as camera-based systems struggle with long detection ranges and adverse lighting and weather conditions. In this work, we introduce HyDRa, a novel camera-radar fusion architecture for diverse 3D perception tasks. Building upon the principles of dense BEV (Bird's Eye View)-based architectures, HyDRa introduces a hybrid fusion approach to combine the strengths of complementary camera and radar features in two distinct representation spaces. Our Height Association Transformer module leverages radar features already in the perspective view to produce more robust and accurate depth predictions. In the BEV, we refine the initial sparse representation by a Radar-weighted Depth Consistency. HyDRa achieves a new state-of-the-art for camera-radar fusion of 64.2 NDS (+1.8) and 58.4 AMOTA (+1.5) on the public nuScenes dataset. Moreover, our new semantically rich and spatially accurate BEV features can be directly converted into a powerful occupancy representation, beating all previous camera-based methods on the Occ3D benchmark by an impressive 3.7 mIoU. Code and models are available at https://github.com/phi-wol/hydra.

Results

TaskDatasetMetricValueModel
Multi-Object Trackingnuscenes Camera-RadarAMOTA0.584HyDRa
Multi-Object TrackingnuScenesAMOTA0.584HyDRa
Object Trackingnuscenes Camera-RadarAMOTA0.584HyDRa
Object TrackingnuScenesAMOTA0.584HyDRa
Prediction Of Occupancy Grid MapsOcc3D-nuScenesmIoU44.4HyDRa R50
Object DetectionView-of-Delft (val)mAP60.9HyDRa
Object DetectionnuScenesNDS0.64HyDRa
Object DetectionnuScenesmAAE0.12HyDRa
Object DetectionnuScenesmAOE0.42HyDRa
Object DetectionnuScenesmAP0.57HyDRa
Object DetectionnuScenesmASE0.25HyDRa
Object DetectionnuScenesmATE0.4HyDRa
Object DetectionnuScenesmAVE0.25HyDRa
Object Detectionnuscenes Camera-RadarNDS64.2HyDRa
Object DetectionTruckScenesNDS22.4HyDRa
Object DetectionTruckScenesmAP12.8HyDRa
3DView-of-Delft (val)mAP60.9HyDRa
3DnuScenesNDS0.64HyDRa
3DnuScenesmAAE0.12HyDRa
3DnuScenesmAOE0.42HyDRa
3DnuScenesmAP0.57HyDRa
3DnuScenesmASE0.25HyDRa
3DnuScenesmATE0.4HyDRa
3DnuScenesmAVE0.25HyDRa
3Dnuscenes Camera-RadarNDS64.2HyDRa
3DTruckScenesNDS22.4HyDRa
3DTruckScenesmAP12.8HyDRa
3D Object DetectionView-of-Delft (val)mAP60.9HyDRa
3D Object DetectionnuScenesNDS0.64HyDRa
3D Object DetectionnuScenesmAAE0.12HyDRa
3D Object DetectionnuScenesmAOE0.42HyDRa
3D Object DetectionnuScenesmAP0.57HyDRa
3D Object DetectionnuScenesmASE0.25HyDRa
3D Object DetectionnuScenesmATE0.4HyDRa
3D Object DetectionnuScenesmAVE0.25HyDRa
3D Object Detectionnuscenes Camera-RadarNDS64.2HyDRa
3D Object DetectionTruckScenesNDS22.4HyDRa
3D Object DetectionTruckScenesmAP12.8HyDRa
3D Multi-Object Trackingnuscenes Camera-RadarAMOTA0.584HyDRa
3D Multi-Object TrackingnuScenesAMOTA0.584HyDRa
2D ClassificationView-of-Delft (val)mAP60.9HyDRa
2D ClassificationnuScenesNDS0.64HyDRa
2D ClassificationnuScenesmAAE0.12HyDRa
2D ClassificationnuScenesmAOE0.42HyDRa
2D ClassificationnuScenesmAP0.57HyDRa
2D ClassificationnuScenesmASE0.25HyDRa
2D ClassificationnuScenesmATE0.4HyDRa
2D ClassificationnuScenesmAVE0.25HyDRa
2D Classificationnuscenes Camera-RadarNDS64.2HyDRa
2D ClassificationTruckScenesNDS22.4HyDRa
2D ClassificationTruckScenesmAP12.8HyDRa
2D Object DetectionView-of-Delft (val)mAP60.9HyDRa
2D Object DetectionnuScenesNDS0.64HyDRa
2D Object DetectionnuScenesmAAE0.12HyDRa
2D Object DetectionnuScenesmAOE0.42HyDRa
2D Object DetectionnuScenesmAP0.57HyDRa
2D Object DetectionnuScenesmASE0.25HyDRa
2D Object DetectionnuScenesmATE0.4HyDRa
2D Object DetectionnuScenesmAVE0.25HyDRa
2D Object Detectionnuscenes Camera-RadarNDS64.2HyDRa
2D Object DetectionTruckScenesNDS22.4HyDRa
2D Object DetectionTruckScenesmAP12.8HyDRa
16kView-of-Delft (val)mAP60.9HyDRa
16knuScenesNDS0.64HyDRa
16knuScenesmAAE0.12HyDRa
16knuScenesmAOE0.42HyDRa
16knuScenesmAP0.57HyDRa
16knuScenesmASE0.25HyDRa
16knuScenesmATE0.4HyDRa
16knuScenesmAVE0.25HyDRa
16knuscenes Camera-RadarNDS64.2HyDRa
16kTruckScenesNDS22.4HyDRa
16kTruckScenesmAP12.8HyDRa

Related Papers

GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17Safeguarding Federated Learning-based Road Condition Classification2025-07-16