TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Auto-Rectify Network for Unsupervised Indoor Depth Estimat...

Auto-Rectify Network for Unsupervised Indoor Depth Estimation

Jia-Wang Bian, Huangying Zhan, Naiyan Wang, Tat-Jun Chin, Chunhua Shen, Ian Reid

2020-06-04Self-Supervised LearningTranslationDepth EstimationMonocular Depth Estimation
PaperPDFCode

Abstract

Single-View depth estimation using the CNNs trained from unlabelled videos has shown significant promise. However, excellent results have mostly been obtained in street-scene driving scenarios, and such methods often fail in other settings, particularly indoor videos taken by handheld devices. In this work, we establish that the complex ego-motions exhibited in handheld settings are a critical obstacle for learning depth. Our fundamental analysis suggests that the rotation behaves as noise during training, as opposed to the translation (baseline) which provides supervision signals. To address the challenge, we propose a data pre-processing method that rectifies training images by removing their relative rotations for effective learning. The significantly improved performance validates our motivation. Towards end-to-end learning without requiring pre-processing, we propose an Auto-Rectify Network with novel loss functions, which can automatically learn to rectify images during training. Consequently, our results outperform the previous unsupervised SOTA method by a large margin on the challenging NYUv2 dataset. We also demonstrate the generalization of our trained model in ScanNet and Make3D, and the universality of our proposed learning method on 7-Scenes and KITTI datasets.

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2Delta < 1.250.82SC-DepthV2
Depth EstimationNYU-Depth V2Delta < 1.25^20.956SC-DepthV2
Depth EstimationNYU-Depth V2RMSE0.532SC-DepthV2
Depth EstimationNYU-Depth V2absolute relative error0.138SC-DepthV2
Depth EstimationNYU-Depth V2log 100.059SC-DepthV2
3DNYU-Depth V2Delta < 1.250.82SC-DepthV2
3DNYU-Depth V2Delta < 1.25^20.956SC-DepthV2
3DNYU-Depth V2RMSE0.532SC-DepthV2
3DNYU-Depth V2absolute relative error0.138SC-DepthV2
3DNYU-Depth V2log 100.059SC-DepthV2

Related Papers

A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16Function-to-Style Guidance of LLMs for Code Translation2025-07-15MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15