TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Deep Ordinal Regression Network for Monocular Depth Estima...

Deep Ordinal Regression Network for Monocular Depth Estimation

Huan Fu, Mingming Gong, Chaohui Wang, Kayhan Batmanghelich, DaCheng Tao

2018-06-06CVPR 2018 6regressionDepth EstimationMonocular Depth Estimation
PaperPDFCodeCodeCodeCodeCode(official)

Abstract

Monocular depth estimation, which plays a crucial role in understanding 3D scene geometry, is an ill-posed problem. Recent methods have gained significant improvement by exploring image-level information and hierarchical features from deep convolutional neural networks (DCNNs). These methods model depth estimation as a regression problem and train the regression networks by minimizing mean squared error, which suffers from slow convergence and unsatisfactory local solutions. Besides, existing depth estimation networks employ repeated spatial pooling operations, resulting in undesirable low-resolution feature maps. To obtain high-resolution depth maps, skip-connections or multi-layer deconvolution networks are required, which complicates network training and consumes much more computations. To eliminate or at least largely reduce these problems, we introduce a spacing-increasing discretization (SID) strategy to discretize depth and recast depth network learning as an ordinal regression problem. By training the network using an ordinary regression loss, our method achieves much higher accuracy and \dd{faster convergence in synch}. Furthermore, we adopt a multi-scale network structure which avoids unnecessary spatial pooling and captures multi-scale information in parallel. The method described in this paper achieves state-of-the-art results on four challenging benchmarks, i.e., KITTI [17], ScanNet [9], Make3D [50], and NYU Depth v2 [42], and win the 1st prize in Robust Vision Challenge 2018. Code has been made available at: https://github.com/hufu6371/DORN.

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2RMS0.509DORN
Depth EstimationNYU-Depth V2RMSE0.509DORN
Depth EstimationKITTI Eigen splitDelta < 1.250.932DORN
Depth EstimationKITTI Eigen splitDelta < 1.25^20.984DORN
Depth EstimationKITTI Eigen splitDelta < 1.25^30.994DORN
Depth EstimationKITTI Eigen splitRMSE2.727DORN
Depth EstimationKITTI Eigen splitRMSE log0.12DORN
Depth EstimationKITTI Eigen splitabsolute relative error0.072DORN
3DNYU-Depth V2RMS0.509DORN
3DNYU-Depth V2RMSE0.509DORN
3DKITTI Eigen splitDelta < 1.250.932DORN
3DKITTI Eigen splitDelta < 1.25^20.984DORN
3DKITTI Eigen splitDelta < 1.25^30.994DORN
3DKITTI Eigen splitRMSE2.727DORN
3DKITTI Eigen splitRMSE log0.12DORN
3DKITTI Eigen splitabsolute relative error0.072DORN

Related Papers

Language Integration in Fine-Tuning Multimodal Large Language Models for Image-Based Regression2025-07-20$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Neural Network-Guided Symbolic Regression for Interpretable Descriptor Discovery in Perovskite Catalysts2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16Second-Order Bounds for [0,1]-Valued Regression via Betting Loss2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16