TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/VA-DepthNet: A Variational Approach to Single Image Depth ...

VA-DepthNet: A Variational Approach to Single Image Depth Prediction

Ce Liu, Suryansh Kumar, Shuhang Gu, Radu Timofte, Luc van Gool

2023-02-13Depth PredictionDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)Code

Abstract

We introduce VA-DepthNet, a simple, effective, and accurate deep neural network approach for the single-image depth prediction (SIDP) problem. The proposed approach advocates using classical first-order variational constraints for this problem. While state-of-the-art deep neural network methods for SIDP learn the scene depth from images in a supervised setting, they often overlook the invaluable invariances and priors in the rigid scene space, such as the regularity of the scene. The paper's main contribution is to reveal the benefit of classical and well-founded variational constraints in the neural network design for the SIDP task. It is shown that imposing first-order variational constraints in the scene space together with popular encoder-decoder-based network architecture design provides excellent results for the supervised SIDP task. The imposed first-order variational constraint makes the network aware of the depth gradient in the scene space, i.e., regularity. The paper demonstrates the usefulness of the proposed approach via extensive evaluation and ablation analysis over several benchmark datasets, such as KITTI, NYU Depth V2, and SUN RGB-D. The VA-DepthNet at test time shows considerable improvements in depth prediction accuracy compared to the prior art and is accurate also at high-frequency regions in the scene space. At the time of writing this paper, our method -- labeled as VA-DepthNet, when tested on the KITTI depth-prediction evaluation set benchmarks, shows state-of-the-art results, and is the top-performing published approach.

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2Delta < 1.250.937VA-DepthNet(SwinV1-L)
Depth EstimationNYU-Depth V2Delta < 1.25^20.992VA-DepthNet(SwinV1-L)
Depth EstimationNYU-Depth V2Delta < 1.25^30.999VA-DepthNet(SwinV1-L)
Depth EstimationNYU-Depth V2RMSE0.304VA-DepthNet(SwinV1-L)
Depth EstimationNYU-Depth V2absolute relative error0.086VA-DepthNet(SwinV1-L)
Depth EstimationNYU-Depth V2log 100.037VA-DepthNet(SwinV1-L)
3DNYU-Depth V2Delta < 1.250.937VA-DepthNet(SwinV1-L)
3DNYU-Depth V2Delta < 1.25^20.992VA-DepthNet(SwinV1-L)
3DNYU-Depth V2Delta < 1.25^30.999VA-DepthNet(SwinV1-L)
3DNYU-Depth V2RMSE0.304VA-DepthNet(SwinV1-L)
3DNYU-Depth V2absolute relative error0.086VA-DepthNet(SwinV1-L)
3DNYU-Depth V2log 100.037VA-DepthNet(SwinV1-L)

Related Papers

$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Cameras as Relative Positional Encoding2025-07-14ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way2025-07-11