TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/M4Depth: Monocular depth estimation for autonomous vehicle...

M4Depth: Monocular depth estimation for autonomous vehicles in unseen environments

Michaël Fonder, Damien Ernst, Marc Van Droogenbroeck

2021-05-20Autonomous VehiclesDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Estimating the distance to objects is crucial for autonomous vehicles when using depth sensors is not possible. In this case, the distance has to be estimated from on-board mounted RGB cameras, which is a complex task especially in environments such as natural outdoor landscapes. In this paper, we present a new method named M4Depth for depth estimation. First, we establish a bijective relationship between depth and the visual disparity of two consecutive frames and show how to exploit it to perform motion-invariant pixel-wise depth estimation. Then, we detail M4Depth which is based on a pyramidal convolutional neural network architecture where each level refines an input disparity map estimate by using two customized cost volumes. We use these cost volumes to leverage the visual spatio-temporal constraints imposed by motion and to make the network robust for varied scenes. We benchmarked our approach both in test and generalization modes on public datasets featuring synthetic camera trajectories recorded in a wide variety of outdoor scenes. Results show that our network outperforms the state of the art on these datasets, while also performing well on a standard depth estimation benchmark. The code of our method is publicly available at https://github.com/michael-fonder/M4Depth.

Results

TaskDatasetMetricValueModel
Depth EstimationMid-Air DatasetAbs Rel0.1425M4Depth-d6 (VMD)
Depth EstimationMid-Air DatasetRMSE8.8641M4Depth-d6 (VMD)
Depth EstimationMid-Air DatasetRMSE log0.24571M4Depth-d6 (VMD)
Depth EstimationMid-Air DatasetSQ Rel3.6798M4Depth-d6 (VMD)
3DMid-Air DatasetAbs Rel0.1425M4Depth-d6 (VMD)
3DMid-Air DatasetRMSE8.8641M4Depth-d6 (VMD)
3DMid-Air DatasetRMSE log0.24571M4Depth-d6 (VMD)
3DMid-Air DatasetSQ Rel3.6798M4Depth-d6 (VMD)

Related Papers

$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Cameras as Relative Positional Encoding2025-07-14ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way2025-07-11