TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/StructDepth: Leveraging the structural regularities for se...

StructDepth: Leveraging the structural regularities for self-supervised indoor depth estimation

Boying Li, Yuan Huang, Zeyu Liu, Danping Zou, Wenxian Yu

2021-08-19ICCV 2021 10Depth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Self-supervised monocular depth estimation has achieved impressive performance on outdoor datasets. Its performance however degrades notably in indoor environments because of the lack of textures. Without rich textures, the photometric consistency is too weak to train a good depth network. Inspired by the early works on indoor modeling, we leverage the structural regularities exhibited in indoor scenes, to train a better depth network. Specifically, we adopt two extra supervisory signals for self-supervised training: 1) the Manhattan normal constraint and 2) the co-planar constraint. The Manhattan normal constraint enforces the major surfaces (the floor, ceiling, and walls) to be aligned with dominant directions. The co-planar constraint states that the 3D points be well fitted by a plane if they are located within the same planar region. To generate the supervisory signals, we adopt two components to classify the major surface normal into dominant directions and detect the planar regions on the fly during training. As the predicted depth becomes more accurate after more training epochs, the supervisory signals also improve and in turn feedback to obtain a better depth model. Through extensive experiments on indoor benchmark datasets, the results show that our network outperforms the state-of-the-art methods. The source code is available at https://github.com/SJTU-ViSYS/StructDepth .

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2 self-supervisedAbsolute relative error (AbsRel)0.142StrutDepth
Depth EstimationNYU-Depth V2 self-supervisedRoot mean square error (RMSE)0.54StrutDepth
Depth EstimationNYU-Depth V2 self-superviseddelta_181.3StrutDepth
Depth EstimationNYU-Depth V2 self-superviseddelta_295.4StrutDepth
Depth EstimationNYU-Depth V2 self-superviseddelta_398.8StrutDepth
3DNYU-Depth V2 self-supervisedAbsolute relative error (AbsRel)0.142StrutDepth
3DNYU-Depth V2 self-supervisedRoot mean square error (RMSE)0.54StrutDepth
3DNYU-Depth V2 self-superviseddelta_181.3StrutDepth
3DNYU-Depth V2 self-superviseddelta_295.4StrutDepth
3DNYU-Depth V2 self-superviseddelta_398.8StrutDepth

Related Papers

$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Cameras as Relative Positional Encoding2025-07-14ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way2025-07-11