TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/End-to-end Learning for Joint Depth and Image Reconstructi...

End-to-end Learning for Joint Depth and Image Reconstruction from Diffracted Rotation

Mazen Mel, Muhammad Siddiqui, Pietro Zanuttigh

2022-04-14DeblurringImage DeblurringImage ReconstructionDepth EstimationMonocular Depth Estimation
PaperPDF

Abstract

Monocular depth estimation is still an open challenge due to the ill-posed nature of the problem at hand. Deep learning based techniques have been extensively studied and proved capable of producing acceptable depth estimation accuracy even if the lack of meaningful and robust depth cues within single RGB input images severally limits their performance. Coded aperture-based methods using phase and amplitude masks encode strong depth cues within 2D images by means of depth-dependent Point Spread Functions (PSFs) at the price of a reduced image quality. In this paper, we propose a novel end-to-end learning approach for depth from diffracted rotation. A phase mask that produces a Rotating Point Spread Function (RPSF) as a function of defocus is jointly optimized with the weights of a depth estimation neural network. To this aim, we introduce a differentiable physical model of the aperture mask and exploit an accurate simulation of the camera imaging pipeline. Our approach requires a significantly less complex model and less training data, yet it is superior to existing methods in the task of monocular depth estimation on indoor benchmarks. In addition, we address the problem of image degradation by incorporating a non-blind and non-uniform image deblurring module to recover the sharp all-in-focus image from its RPSF-blurred counterpart.

Results

TaskDatasetMetricValueModel
Depth EstimationSUN-RGBDDelta < 1.250.937RPSF
Depth EstimationSUN-RGBDDelta < 1.25^20.981RPSF
Depth EstimationSUN-RGBDDelta < 1.25^30.992RPSF
Depth EstimationSUN-RGBDRMSE0.335RPSF
Depth EstimationSUN-RGBDabsolute relative error0.114RPSF
Depth EstimationSUN-RGBDlog 100.034RPSF
3DSUN-RGBDDelta < 1.250.937RPSF
3DSUN-RGBDDelta < 1.25^20.981RPSF
3DSUN-RGBDDelta < 1.25^30.992RPSF
3DSUN-RGBDRMSE0.335RPSF
3DSUN-RGBDabsolute relative error0.114RPSF
3DSUN-RGBDlog 100.034RPSF

Related Papers

$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16The model is the message: Lightweight convolutional autoencoders applied to noisy imaging data for planetary science and astrobiology2025-07-153D Magnetic Inverse Routine for Single-Segment Magnetic Field Images2025-07-15MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15Towards Depth Foundation Model: Recent Trends in Vision-Based Depth Estimation2025-07-15