TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Adaptive Fusion of Single-View and Multi-View Depth for Au...

Adaptive Fusion of Single-View and Multi-View Depth for Autonomous Driving

Junda Cheng, Wei Yin, Kaixuan Wang, Xiaozhi Chen, Shijie Wang, Xin Yang

2024-03-12CVPR 2024 1Autonomous DrivingDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Multi-view depth estimation has achieved impressive performance over various benchmarks. However, almost all current multi-view systems rely on given ideal camera poses, which are unavailable in many real-world scenarios, such as autonomous driving. In this work, we propose a new robustness benchmark to evaluate the depth estimation system under various noisy pose settings. Surprisingly, we find current multi-view depth estimation methods or single-view and multi-view fusion methods will fail when given noisy pose settings. To address this challenge, we propose a single-view and multi-view fused depth estimation system, which adaptively integrates high-confident multi-view and single-view results for both robust and accurate depth estimations. The adaptive fusion module performs fusion by dynamically selecting high-confidence regions between two branches based on a wrapping confidence map. Thus, the system tends to choose the more reliable branch when facing textureless scenes, inaccurate calibration, dynamic objects, and other degradation or challenging conditions. Our method outperforms state-of-the-art multi-view and fusion methods under robustness testing. Furthermore, we achieve state-of-the-art performance on challenging benchmarks (KITTI and DDAD) when given accurate pose estimations. Project website: https://github.com/Junda24/AFNet/.

Results

TaskDatasetMetricValueModel
Depth EstimationKITTI Eigen splitDelta < 1.250.98AFNet
Depth EstimationKITTI Eigen splitDelta < 1.25^20.997AFNet
Depth EstimationKITTI Eigen splitDelta < 1.25^30.999AFNet
Depth EstimationKITTI Eigen splitRMSE1.712AFNet
Depth EstimationKITTI Eigen splitRMSE log0.069AFNet
Depth EstimationKITTI Eigen splitSq Rel0.132AFNet
Depth EstimationKITTI Eigen splitabsolute relative error0.044AFNet
Depth EstimationDDADRMSE4.6AFNet
Depth EstimationDDADRMSE log0.154AFNet
Depth EstimationDDADSq Rel0.979AFNet
Depth EstimationDDADabsolute relative error0.088AFNet
3DKITTI Eigen splitDelta < 1.250.98AFNet
3DKITTI Eigen splitDelta < 1.25^20.997AFNet
3DKITTI Eigen splitDelta < 1.25^30.999AFNet
3DKITTI Eigen splitRMSE1.712AFNet
3DKITTI Eigen splitRMSE log0.069AFNet
3DKITTI Eigen splitSq Rel0.132AFNet
3DKITTI Eigen splitabsolute relative error0.044AFNet
3DDDADRMSE4.6AFNet
3DDDADRMSE log0.154AFNet
3DDDADSq Rel0.979AFNet
3DDDADabsolute relative error0.088AFNet

Related Papers

GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17