TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/LXL: LiDAR Excluded Lean 3D Object Detection with 4D Imagi...

LXL: LiDAR Excluded Lean 3D Object Detection with 4D Imaging Radar and Camera Fusion

Weiyi Xiong, Jianan Liu, Tao Huang, Qing-Long Han, Yuxuan Xia, Bing Zhu

2023-07-033D Object Detection (RoI)Depth PredictionAutonomous DrivingDepth Estimationobject-detection3D Object DetectionObject Detection
PaperPDF

Abstract

As an emerging technology and a relatively affordable device, the 4D imaging radar has already been confirmed effective in performing 3D object detection in autonomous driving. Nevertheless, the sparsity and noisiness of 4D radar point clouds hinder further performance improvement, and in-depth studies about its fusion with other modalities are lacking. On the other hand, as a new image view transformation strategy, "sampling" has been applied in a few image-based detectors and shown to outperform the widely applied "depth-based splatting" proposed in Lift-Splat-Shoot (LSS), even without image depth prediction. However, the potential of "sampling" is not fully unleashed. This paper investigates the "sampling" view transformation strategy on the camera and 4D imaging radar fusion-based 3D object detection. LiDAR Excluded Lean (LXL) model, predicted image depth distribution maps and radar 3D occupancy grids are generated from image perspective view (PV) features and radar bird's eye view (BEV) features, respectively. They are sent to the core of LXL, called "radar occupancy-assisted depth-based sampling", to aid image view transformation. We demonstrated that more accurate view transformation can be performed by introducing image depths and radar information to enhance the "sampling" strategy. Experiments on VoD and TJ4DRadSet datasets show that the proposed method outperforms the state-of-the-art 3D object detection methods by a significant margin without bells and whistles. Ablation studies demonstrate that our method performs the best among different enhancement settings.

Results

TaskDatasetMetricValueModel
Object DetectionView-of-Delft (val)mAP56.31LXL
Object DetectionView-of-Delft (val)mAP49.3BEVFusion R
Object DetectionView-of-Delft (val)mAP49FUTR3D R
3DView-of-Delft (val)mAP56.31LXL
3DView-of-Delft (val)mAP49.3BEVFusion R
3DView-of-Delft (val)mAP49FUTR3D R
3D Object DetectionView-of-Delft (val)mAP56.31LXL
3D Object DetectionView-of-Delft (val)mAP49.3BEVFusion R
3D Object DetectionView-of-Delft (val)mAP49FUTR3D R
2D ClassificationView-of-Delft (val)mAP56.31LXL
2D ClassificationView-of-Delft (val)mAP49.3BEVFusion R
2D ClassificationView-of-Delft (val)mAP49FUTR3D R
2D Object DetectionView-of-Delft (val)mAP56.31LXL
2D Object DetectionView-of-Delft (val)mAP49.3BEVFusion R
2D Object DetectionView-of-Delft (val)mAP49FUTR3D R
16kView-of-Delft (val)mAP56.31LXL
16kView-of-Delft (val)mAP49.3BEVFusion R
16kView-of-Delft (val)mAP49FUTR3D R

Related Papers

GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17