TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/P$^{2}$Net: Patch-match and Plane-regularization for Unsup...

P$^{2}$Net: Patch-match and Plane-regularization for Unsupervised Indoor Depth Estimation

Zehao Yu, Lei Jin, Shenghua Gao

2020-07-15Depth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

This paper tackles the unsupervised depth estimation task in indoor environments. The task is extremely challenging because of the vast areas of non-texture regions in these scenes. These areas could overwhelm the optimization process in the commonly used unsupervised depth estimation framework proposed for outdoor environments. However, even when those regions are masked out, the performance is still unsatisfactory. In this paper, we argue that the poor performance suffers from the non-discriminative point-based matching. To this end, we propose P$^2$Net. We first extract points with large local gradients and adopt patches centered at each point as its representation. Multiview consistency loss is then defined over patches. This operation significantly improves the robustness of the network training. Furthermore, because those textureless regions in indoor scenes (e.g., wall, floor, roof, \etc) usually correspond to planar regions, we propose to leverage superpixels as a plane prior. We enforce the predicted depth to be well fitted by a plane within each superpixel. Extensive experiments on NYUv2 and ScanNet show that our P$^2$Net outperforms existing approaches by a large margin. Code is available at \url{https://github.com/svip-lab/Indoor-SfMLearner}.

Results

TaskDatasetMetricValueModel
Depth EstimationNYU-Depth V2 self-supervisedAbsolute relative error (AbsRel)0.147P2Net+PP
Depth EstimationNYU-Depth V2 self-supervisedRoot mean square error (RMSE)0.553P2Net+PP
Depth EstimationNYU-Depth V2 self-superviseddelta_180.4P2Net+PP
Depth EstimationNYU-Depth V2 self-superviseddelta_295.2P2Net+PP
Depth EstimationNYU-Depth V2 self-superviseddelta_398.7P2Net+PP
3DNYU-Depth V2 self-supervisedAbsolute relative error (AbsRel)0.147P2Net+PP
3DNYU-Depth V2 self-supervisedRoot mean square error (RMSE)0.553P2Net+PP
3DNYU-Depth V2 self-superviseddelta_180.4P2Net+PP
3DNYU-Depth V2 self-superviseddelta_295.2P2Net+PP
3DNYU-Depth V2 self-superviseddelta_398.7P2Net+PP

Related Papers

$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Cameras as Relative Positional Encoding2025-07-14ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way2025-07-11