TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/DS-Depth: Dynamic and Static Depth Estimation via a Fusion...

DS-Depth: Dynamic and Static Depth Estimation via a Fusion Cost Volume

Xingyu Miao, Yang Bai, Haoran Duan, Yawen Huang, Fan Wan, Xinxing Xu, Yang Long, Yefeng Zheng

2023-08-14Optical Flow EstimationUnsupervised Monocular Depth EstimationDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Self-supervised monocular depth estimation methods typically rely on the reprojection error to capture geometric relationships between successive frames in static environments. However, this assumption does not hold in dynamic objects in scenarios, leading to errors during the view synthesis stage, such as feature mismatch and occlusion, which can significantly reduce the accuracy of the generated depth maps. To address this problem, we propose a novel dynamic cost volume that exploits residual optical flow to describe moving objects, improving incorrectly occluded regions in static cost volumes used in previous work. Nevertheless, the dynamic cost volume inevitably generates extra occlusions and noise, thus we alleviate this by designing a fusion module that makes static and dynamic cost volumes compensate for each other. In other words, occlusion from the static volume is refined by the dynamic volume, and incorrect information from the dynamic volume is eliminated by the static volume. Furthermore, we propose a pyramid distillation loss to reduce photometric error inaccuracy at low resolutions and an adaptive photometric error loss to alleviate the flow direction of the large gradient in the occlusion regions. We conducted extensive experiments on the KITTI and Cityscapes datasets, and the results demonstrate that our model outperforms previously published baselines for self-supervised monocular depth estimation.

Results

TaskDatasetMetricValueModel
Depth EstimationKITTI Eigen split unsupervisedDelta < 1.250.905DS-Depth
Depth EstimationKITTI Eigen split unsupervisedDelta < 1.25^20.966DS-Depth
Depth EstimationKITTI Eigen split unsupervisedDelta < 1.25^30.984DS-Depth
Depth EstimationKITTI Eigen split unsupervisedRMSE4.329DS-Depth
Depth EstimationKITTI Eigen split unsupervisedRMSE log0.173DS-Depth
Depth EstimationKITTI Eigen split unsupervisedSq Rel0.698DS-Depth
Depth EstimationKITTI Eigen split unsupervisedabsolute relative error0.095DS-Depth
3DKITTI Eigen split unsupervisedDelta < 1.250.905DS-Depth
3DKITTI Eigen split unsupervisedDelta < 1.25^20.966DS-Depth
3DKITTI Eigen split unsupervisedDelta < 1.25^30.984DS-Depth
3DKITTI Eigen split unsupervisedRMSE4.329DS-Depth
3DKITTI Eigen split unsupervisedRMSE log0.173DS-Depth
3DKITTI Eigen split unsupervisedSq Rel0.698DS-Depth
3DKITTI Eigen split unsupervisedabsolute relative error0.095DS-Depth

Related Papers

Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15Cameras as Relative Positional Encoding2025-07-14