TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/P3Depth: Monocular Depth Estimation with a Piecewise Plana...

P3Depth: Monocular Depth Estimation with a Piecewise Planarity Prior

Vaishakh Patil, Christos Sakaridis, Alexander Liniger, Luc van Gool

2022-04-05CVPR 2022 1Scene UnderstandingDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Monocular depth estimation is vital for scene understanding and downstream tasks. We focus on the supervised setup, in which ground-truth depth is available only at training time. Based on knowledge about the high regularity of real 3D scenes, we propose a method that learns to selectively leverage information from coplanar pixels to improve the predicted depth. In particular, we introduce a piecewise planarity prior which states that for each pixel, there is a seed pixel which shares the same planar 3D surface with the former. Motivated by this prior, we design a network with two heads. The first head outputs pixel-level plane coefficients, while the second one outputs a dense offset vector field that identifies the positions of seed pixels. The plane coefficients of seed pixels are then used to predict depth at each position. The resulting prediction is adaptively fused with the initial prediction from the first head via a learned confidence to account for potential deviations from precise local planarity. The entire architecture is trained end-to-end thanks to the differentiability of the proposed modules and it learns to predict regular depth maps, with sharp edges at occlusion boundaries. An extensive evaluation of our method shows that we set the new state of the art in supervised monocular depth estimation, surpassing prior methods on NYU Depth-v2 and on the Garg split of KITTI. Our method delivers depth maps that yield plausible 3D reconstructions of the input scenes. Code is available at: https://github.com/SysCV/P3Depth

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2RMS0.356P3Depth
Depth EstimationNYU-Depth V2Delta < 1.250.898P3Depth
Depth EstimationNYU-Depth V2Delta < 1.25^20.981P3Depth
Depth EstimationNYU-Depth V2Delta < 1.25^30.996P3Depth
Depth EstimationNYU-Depth V2RMSE0.356P3Depth
Depth EstimationNYU-Depth V2absolute relative error0.104P3Depth
Depth EstimationNYU-Depth V2log 100.043P3Depth
3DNYU-Depth V2RMS0.356P3Depth
3DNYU-Depth V2Delta < 1.250.898P3Depth
3DNYU-Depth V2Delta < 1.25^20.981P3Depth
3DNYU-Depth V2Delta < 1.25^30.996P3Depth
3DNYU-Depth V2RMSE0.356P3Depth
3DNYU-Depth V2absolute relative error0.104P3Depth
3DNYU-Depth V2log 100.043P3Depth

Related Papers

Advancing Complex Wide-Area Scene Understanding with Hierarchical Coresets Selection2025-07-17Argus: Leveraging Multiview Images for Improved 3-D Scene Understanding With Large Language Models2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Learning to Tune Like an Expert: Interpretable and Scene-Aware Navigation via MLLM Reasoning and CVAE-Based Adaptation2025-07-15