TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning to Recover 3D Scene Shape from a Single Image

Learning to Recover 3D Scene Shape from a Single Image

Wei Yin, Jianming Zhang, Oliver Wang, Simon Niklaus, Long Mai, Simon Chen, Chunhua Shen

2020-12-17CVPR 2021 1Indoor Monocular Depth Estimation3D Scene ReconstructionDepth PredictionDepth EstimationSingle-View 3D ReconstructionMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Despite significant progress in monocular depth estimation in the wild, recent state-of-the-art methods cannot be used to recover accurate 3D scene shape due to an unknown depth shift induced by shift-invariant reconstruction losses used in mixed-data depth prediction training, and possible unknown camera focal length. We investigate this problem in detail, and propose a two-stage framework that first predicts depth up to an unknown scale and shift from a single monocular image, and then use 3D point cloud encoders to predict the missing depth shift and focal length that allow us to recover a realistic 3D scene shape. In addition, we propose an image-level normalized regression loss and a normal-based geometry loss to enhance depth prediction models trained on mixed datasets. We test our depth model on nine unseen datasets and achieve state-of-the-art performance on zero-shot dataset generalization. Code is available at: https://git.io/Depth

Results

TaskDatasetMetricValueModel
Depth EstimationScanNetV2absolute relative error0.095LeReS
Depth EstimationDIODEDelta < 1.250.234LeRes
Depth EstimationNYU-Depth V2Delta < 1.250.916LeReS
Depth EstimationNYU-Depth V2absolute relative error0.09LeReS
Depth EstimationETH3DDelta < 1.250.0777LeReS
Depth EstimationETH3Dabsolute relative error0.0171LeReS
Depth EstimationKITTI Eigen splitDelta < 1.250.784LeReS
Depth EstimationKITTI Eigen splitabsolute relative error0.149LeReS
Depth EstimationDIODEDelta < 1.25^30.9LeReS
3DScanNetV2absolute relative error0.095LeReS
3DDIODEDelta < 1.250.234LeRes
3DNYU-Depth V2Delta < 1.250.916LeReS
3DNYU-Depth V2absolute relative error0.09LeReS
3DETH3DDelta < 1.250.0777LeReS
3DETH3Dabsolute relative error0.0171LeReS
3DKITTI Eigen splitDelta < 1.250.784LeReS
3DKITTI Eigen splitabsolute relative error0.149LeReS
3DDIODEDelta < 1.25^30.9LeReS

Related Papers

$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Physically Based Neural LiDAR Resimulation2025-07-15MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Cameras as Relative Positional Encoding2025-07-14