TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Unsupervised Depth Completion from Visual Inertial Odometry

Unsupervised Depth Completion from Visual Inertial Odometry

Alex Wong, Xiaohan Fei, Stephanie Tsuei, Stefano Soatto

2019-05-15Depth Completion
PaperPDFCode(official)Code

Abstract

We describe a method to infer dense depth from camera motion and sparse depth as estimated using a visual-inertial odometry system. Unlike other scenarios using point clouds from lidar or structured light sensors, we have few hundreds to few thousand points, insufficient to inform the topology of the scene. Our method first constructs a piecewise planar scaffolding of the scene, and then uses it to infer dense depth using the image along with the sparse points. We use a predictive cross-modal criterion, akin to `self-supervision,' measuring photometric consistency across time, forward-backward pose consistency, and geometric compatibility with the sparse point cloud. We also launch the first visual-inertial + depth dataset, which we hope will foster additional exploration into combining the complementary strengths of visual and inertial sensors. To compare our method to prior work, we adopt the unsupervised KITTI depth completion benchmark, and show state-of-the-art performance on it. Code available at: https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry.

Results

TaskDatasetMetricValueModel
Depth CompletionKITTI Depth CompletionMAE299.41VOICED
Depth CompletionKITTI Depth CompletionRMSE1169.97VOICED
Depth CompletionKITTI Depth CompletionRuntime [ms]20VOICED
Depth CompletionKITTI Depth CompletioniMAE1.2VOICED
Depth CompletionKITTI Depth CompletioniRMSE3.56VOICED
Depth CompletionVOIDMAE85.05VOICED
Depth CompletionVOIDRMSE169.79VOICED
Depth CompletionVOIDiMAE48.92VOICED
Depth CompletionVOIDiRMSE104.02VOICED

Related Papers

PacGDC: Label-Efficient Generalizable Depth Completion with Projection Ambiguity and Consistency2025-07-10DidSee: Diffusion-Based Depth Completion for Material-Agnostic Robotic Perception and Manipulation2025-06-26DCIRNet: Depth Completion with Iterative Refinement for Dexterous Grasping of Transparent and Reflective Objects2025-06-11SR3D: Unleashing Single-view 3D Reconstruction for Transparent and Specular Object Grasping2025-05-30HTMNet: A Hybrid Network with Transformer-Mamba Bottleneck Multimodal Fusion for Transparent and Reflective Objects Depth Completion2025-05-27BadDepth: Backdoor Attacks Against Monocular Depth Estimation in the Physical World2025-05-22Event-Driven Dynamic Scene Depth Completion2025-05-19Depth Anything with Any Prior2025-05-15