TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/PlaneDepth: Self-supervised Depth Estimation via Orthogona...

PlaneDepth: Self-supervised Depth Estimation via Orthogonal Planes

Ruoyu Wang, Zehao Yu, Shenghua Gao

2022-10-04CVPR 2023 1Data AugmentationAutonomous DrivingDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Multiple near frontal-parallel planes based depth representation demonstrated impressive results in self-supervised monocular depth estimation (MDE). Whereas, such a representation would cause the discontinuity of the ground as it is perpendicular to the frontal-parallel planes, which is detrimental to the identification of drivable space in autonomous driving. In this paper, we propose the PlaneDepth, a novel orthogonal planes based presentation, including vertical planes and ground planes. PlaneDepth estimates the depth distribution using a Laplacian Mixture Model based on orthogonal planes for an input image. These planes are used to synthesize a reference view to provide the self-supervision signal. Further, we find that the widely used resizing and cropping data augmentation breaks the orthogonality assumptions, leading to inferior plane predictions. We address this problem by explicitly constructing the resizing cropping transformation to rectify the predefined planes and predicted camera pose. Moreover, we propose an augmented self-distillation loss supervised with a bilateral occlusion mask to boost the robustness of orthogonal planes representation for occlusions. Thanks to our orthogonal planes representation, we can extract the ground plane in an unsupervised manner, which is important for autonomous driving. Extensive experiments on the KITTI dataset demonstrate the effectiveness and efficiency of our method. The code is available at https://github.com/svip-lab/PlaneDepth.

Results

TaskDatasetMetricValueModel
Depth EstimationKITTI Eigen split unsupervisedDelta < 1.250.911PlaneDepth (S + 1280x384)
Depth EstimationKITTI Eigen split unsupervisedDelta < 1.25^20.968PlaneDepth (S + 1280x384)
Depth EstimationKITTI Eigen split unsupervisedDelta < 1.25^30.984PlaneDepth (S + 1280x384)
Depth EstimationKITTI Eigen split unsupervisedRMSE3.981PlaneDepth (S + 1280x384)
Depth EstimationKITTI Eigen split unsupervisedRMSE log0.169PlaneDepth (S + 1280x384)
Depth EstimationKITTI Eigen split unsupervisedSq Rel0.549PlaneDepth (S + 1280x384)
Depth EstimationKITTI Eigen split unsupervisedabsolute relative error0.084PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedDelta < 1.250.911PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedDelta < 1.25^20.968PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedDelta < 1.25^30.984PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedRMSE3.981PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedRMSE log0.169PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedSq Rel0.549PlaneDepth (S + 1280x384)
3DKITTI Eigen split unsupervisedabsolute relative error0.084PlaneDepth (S + 1280x384)

Related Papers

GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17