TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/3D Dual-Fusion: Dual-Domain Dual-Query Camera-LiDAR Fusion...

3D Dual-Fusion: Dual-Domain Dual-Query Camera-LiDAR Fusion for 3D Object Detection

Yecheol Kim, Konyul Park, Minwook Kim, Dongsuk Kum, Jun Won Choi

2022-11-24object-detectionRobust 3D Object Detection3D Object DetectionObject Detection
PaperPDFCode(official)

Abstract

Fusing data from cameras and LiDAR sensors is an essential technique to achieve robust 3D object detection. One key challenge in camera-LiDAR fusion involves mitigating the large domain gap between the two sensors in terms of coordinates and data distribution when fusing their features. In this paper, we propose a novel camera-LiDAR fusion architecture called, 3D Dual-Fusion, which is designed to mitigate the gap between the feature representations of camera and LiDAR data. The proposed method fuses the features of the camera-view and 3D voxel-view domain and models their interactions through deformable attention. We redesign the transformer fusion encoder to aggregate the information from the two domains. Two major changes include 1) dual query-based deformable attention to fuse the dual-domain features interactively and 2) 3D local self-attention to encode the voxel-domain queries prior to dual-query decoding. The results of an experimental evaluation show that the proposed camera-LiDAR fusion architecture achieved competitive performance on the KITTI and nuScenes datasets, with state-of-the-art performances in some 3D object detection benchmarks categories.

Results

TaskDatasetMetricValueModel
Object DetectionnuScenesNDS0.733D Dual-Fusion_T
Object DetectionnuScenesmAAE0.133D Dual-Fusion_T
Object DetectionnuScenesmAOE0.333D Dual-Fusion_T
Object DetectionnuScenesmAP0.713D Dual-Fusion_T
Object DetectionnuScenesmASE0.243D Dual-Fusion_T
Object DetectionnuScenesmATE0.263D Dual-Fusion_T
Object DetectionnuScenesmAVE0.273D Dual-Fusion_T
Object DetectionnuScenesNDS0.733D Dual-Fusion_T
Object DetectionnuScenesmAAE0.133D Dual-Fusion_T
Object DetectionnuScenesmAOE0.333D Dual-Fusion_T
Object DetectionnuScenesmAP0.713D Dual-Fusion_T
Object DetectionnuScenesmASE0.243D Dual-Fusion_T
Object DetectionnuScenesmATE0.263D Dual-Fusion_T
Object DetectionnuScenesmAVE0.273D Dual-Fusion_T
3DnuScenesNDS0.733D Dual-Fusion_T
3DnuScenesmAAE0.133D Dual-Fusion_T
3DnuScenesmAOE0.333D Dual-Fusion_T
3DnuScenesmAP0.713D Dual-Fusion_T
3DnuScenesmASE0.243D Dual-Fusion_T
3DnuScenesmATE0.263D Dual-Fusion_T
3DnuScenesmAVE0.273D Dual-Fusion_T
3DnuScenesNDS0.733D Dual-Fusion_T
3DnuScenesmAAE0.133D Dual-Fusion_T
3DnuScenesmAOE0.333D Dual-Fusion_T
3DnuScenesmAP0.713D Dual-Fusion_T
3DnuScenesmASE0.243D Dual-Fusion_T
3DnuScenesmATE0.263D Dual-Fusion_T
3DnuScenesmAVE0.273D Dual-Fusion_T
3D Object DetectionnuScenesNDS0.733D Dual-Fusion_T
3D Object DetectionnuScenesmAAE0.133D Dual-Fusion_T
3D Object DetectionnuScenesmAOE0.333D Dual-Fusion_T
3D Object DetectionnuScenesmAP0.713D Dual-Fusion_T
3D Object DetectionnuScenesmASE0.243D Dual-Fusion_T
3D Object DetectionnuScenesmATE0.263D Dual-Fusion_T
3D Object DetectionnuScenesmAVE0.273D Dual-Fusion_T
3D Object DetectionnuScenesNDS0.733D Dual-Fusion_T
3D Object DetectionnuScenesmAAE0.133D Dual-Fusion_T
3D Object DetectionnuScenesmAOE0.333D Dual-Fusion_T
3D Object DetectionnuScenesmAP0.713D Dual-Fusion_T
3D Object DetectionnuScenesmASE0.243D Dual-Fusion_T
3D Object DetectionnuScenesmATE0.263D Dual-Fusion_T
3D Object DetectionnuScenesmAVE0.273D Dual-Fusion_T
2D ClassificationnuScenesNDS0.733D Dual-Fusion_T
2D ClassificationnuScenesmAAE0.133D Dual-Fusion_T
2D ClassificationnuScenesmAOE0.333D Dual-Fusion_T
2D ClassificationnuScenesmAP0.713D Dual-Fusion_T
2D ClassificationnuScenesmASE0.243D Dual-Fusion_T
2D ClassificationnuScenesmATE0.263D Dual-Fusion_T
2D ClassificationnuScenesmAVE0.273D Dual-Fusion_T
2D ClassificationnuScenesNDS0.733D Dual-Fusion_T
2D ClassificationnuScenesmAAE0.133D Dual-Fusion_T
2D ClassificationnuScenesmAOE0.333D Dual-Fusion_T
2D ClassificationnuScenesmAP0.713D Dual-Fusion_T
2D ClassificationnuScenesmASE0.243D Dual-Fusion_T
2D ClassificationnuScenesmATE0.263D Dual-Fusion_T
2D ClassificationnuScenesmAVE0.273D Dual-Fusion_T
2D Object DetectionnuScenesNDS0.733D Dual-Fusion_T
2D Object DetectionnuScenesmAAE0.133D Dual-Fusion_T
2D Object DetectionnuScenesmAOE0.333D Dual-Fusion_T
2D Object DetectionnuScenesmAP0.713D Dual-Fusion_T
2D Object DetectionnuScenesmASE0.243D Dual-Fusion_T
2D Object DetectionnuScenesmATE0.263D Dual-Fusion_T
2D Object DetectionnuScenesmAVE0.273D Dual-Fusion_T
2D Object DetectionnuScenesNDS0.733D Dual-Fusion_T
2D Object DetectionnuScenesmAAE0.133D Dual-Fusion_T
2D Object DetectionnuScenesmAOE0.333D Dual-Fusion_T
2D Object DetectionnuScenesmAP0.713D Dual-Fusion_T
2D Object DetectionnuScenesmASE0.243D Dual-Fusion_T
2D Object DetectionnuScenesmATE0.263D Dual-Fusion_T
2D Object DetectionnuScenesmAVE0.273D Dual-Fusion_T
16knuScenesNDS0.733D Dual-Fusion_T
16knuScenesmAAE0.133D Dual-Fusion_T
16knuScenesmAOE0.333D Dual-Fusion_T
16knuScenesmAP0.713D Dual-Fusion_T
16knuScenesmASE0.243D Dual-Fusion_T
16knuScenesmATE0.263D Dual-Fusion_T
16knuScenesmAVE0.273D Dual-Fusion_T
16knuScenesNDS0.733D Dual-Fusion_T
16knuScenesmAAE0.133D Dual-Fusion_T
16knuScenesmAOE0.333D Dual-Fusion_T
16knuScenesmAP0.713D Dual-Fusion_T
16knuScenesmASE0.243D Dual-Fusion_T
16knuScenesmATE0.263D Dual-Fusion_T
16knuScenesmAVE0.273D Dual-Fusion_T

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17Decoupled PROB: Decoupled Query Initialization Tasks and Objectness-Class Learning for Open World Object Detection2025-07-17Dual LiDAR-Based Traffic Movement Count Estimation at a Signalized Intersection: Deployment, Data Collection, and Preliminary Analysis2025-07-17Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Tomato Multi-Angle Multi-Pose Dataset for Fine-Grained Phenotyping2025-07-15ECORE: Energy-Conscious Optimized Routing for Deep Learning Models at the Edge2025-07-08Beyond One Shot, Beyond One Perspective: Cross-View and Long-Horizon Distillation for Better LiDAR Representations2025-07-07