TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Deeper into Self-Supervised Monocular Indoor Depth Estimat...

Deeper into Self-Supervised Monocular Indoor Depth Estimation

Chao Fan, Zhenyu Yin, Yue Li, Feiqing Zhang

2023-12-03Self-Supervised LearningSSIMmotion predictionUnsupervised Monocular Depth EstimationDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Monocular depth estimation using Convolutional Neural Networks (CNNs) has shown impressive performance in outdoor driving scenes. However, self-supervised learning of indoor depth from monocular sequences is quite challenging for researchers because of the following two main reasons. One is the large areas of low-texture regions and the other is the complex ego-motion on indoor training datasets. In this work, our proposed method, named IndoorDepth, consists of two innovations. In particular, we first propose a novel photometric loss with improved structural similarity (SSIM) function to tackle the challenge from low-texture regions. Moreover, in order to further mitigate the issue of inaccurate ego-motion prediction, multiple photometric losses at different stages are used to train a deeper pose network with two residual pose blocks. Subsequent ablation study can validate the effectiveness of each new idea. Experiments on the NYUv2 benchmark demonstrate that our IndoorDepth outperforms the previous state-of-the-art methods by a large margin. In addition, we also validate the generalization ability of our method on ScanNet dataset. Code is availabe at https://github.com/fcntes/IndoorDepth.

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2 self-supervisedAbsolute relative error (AbsRel)0.126IndoorDepth
Depth EstimationNYU-Depth V2 self-supervisedRoot mean square error (RMSE)0.494IndoorDepth
Depth EstimationNYU-Depth V2 self-superviseddelta_184.5IndoorDepth
Depth EstimationNYU-Depth V2 self-superviseddelta_296.5IndoorDepth
Depth EstimationNYU-Depth V2 self-superviseddelta_399.1IndoorDepth
3DNYU-Depth V2 self-supervisedAbsolute relative error (AbsRel)0.126IndoorDepth
3DNYU-Depth V2 self-supervisedRoot mean square error (RMSE)0.494IndoorDepth
3DNYU-Depth V2 self-superviseddelta_184.5IndoorDepth
3DNYU-Depth V2 self-superviseddelta_296.5IndoorDepth
3DNYU-Depth V2 self-superviseddelta_399.1IndoorDepth

Related Papers

A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17SpectraLift: Physics-Guided Spectral-Inversion Network for Self-Supervised Hyperspectral Image Super-Resolution2025-07-17fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16COLI: A Hierarchical Efficient Compressor for Large Images2025-07-15