TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/GCNDepth: Self-supervised Monocular Depth Estimation based...

GCNDepth: Self-supervised Monocular Depth Estimation based on Graph Convolutional Network

Armin Masoumian, Hatem A. Rashwan, Saddam Abdulwahab, Julian Cristiano, Domenec Puig

2021-12-13Depth Prediction3D ReconstructionDepth EstimationMonocular Depth Estimation
PaperPDFCode(official)

Abstract

Depth estimation is a challenging task of 3D reconstruction to enhance the accuracy sensing of environment awareness. This work brings a new solution with a set of improvements, which increase the quantitative and qualitative understanding of depth maps compared to existing methods. Recently, a convolutional neural network (CNN) has demonstrated its extraordinary ability in estimating depth maps from monocular videos. However, traditional CNN does not support topological structure and they can work only on regular image regions with determined size and weights. On the other hand, graph convolutional networks (GCN) can handle the convolution on non-Euclidean data and it can be applied to irregular image regions within a topological structure. Therefore, in this work in order to preserve object geometric appearances and distributions, we aim at exploiting GCN for a self-supervised depth estimation model. Our model consists of two parallel auto-encoder networks: the first is an auto-encoder that will depend on ResNet-50 and extract the feature from the input image and on multi-scale GCN to estimate the depth map. In turn, the second network will be used to estimate the ego-motion vector (i.e., 3D pose) between two consecutive frames based on ResNet-18. Both the estimated 3D pose and depth map will be used for constructing a target image. A combination of loss functions related to photometric, projection, and smoothness is used to cope with bad depth prediction and preserve the discontinuities of the objects. In particular, our method provided comparable and promising results with a high prediction accuracy of 89% on the publicly KITTI and Make3D datasets along with a reduction of 40% in the number of trainable parameters compared to the state of the art solutions. The source code is publicly available at https://github.com/ArminMasoumian/GCNDepth.git

Results

TaskDatasetMetricValueModel
Depth EstimationKITTIabsolute relative error0.104GCNDepth
Depth EstimationKITTI Eigen splitDelta < 1.250.888GCNDepth
Depth EstimationKITTI Eigen splitDelta < 1.25^20.965GCNDepth
Depth EstimationKITTI Eigen splitDelta < 1.25^30.984GCNDepth
Depth EstimationKITTI Eigen splitRMSE4.494GCNDepth
Depth EstimationKITTI Eigen splitRMSE log0.181GCNDepth
Depth EstimationKITTI Eigen splitabsolute relative error0.104GCNDepth
Depth EstimationMake3DAbs Rel0.424GCNDepth
Depth EstimationMake3DRMSE6.757GCNDepth
Depth EstimationMake3DSq Rel3.075GCNDepth
Depth EstimationKITTI Eigen split unsupervisedRMSE4.494GCNDepth
Depth EstimationKITTI Eigen split unsupervisedRMSE log0.181GCNDepth
Depth EstimationKITTI Eigen split unsupervisedSq Rel0.72GCNDepth
Depth EstimationKITTI Eigen split unsupervisedabsolute relative error0.104GCNDepth
3DKITTIabsolute relative error0.104GCNDepth
3DKITTI Eigen splitDelta < 1.250.888GCNDepth
3DKITTI Eigen splitDelta < 1.25^20.965GCNDepth
3DKITTI Eigen splitDelta < 1.25^30.984GCNDepth
3DKITTI Eigen splitRMSE4.494GCNDepth
3DKITTI Eigen splitRMSE log0.181GCNDepth
3DKITTI Eigen splitabsolute relative error0.104GCNDepth
3DMake3DAbs Rel0.424GCNDepth
3DMake3DRMSE6.757GCNDepth
3DMake3DSq Rel3.075GCNDepth
3DKITTI Eigen split unsupervisedRMSE4.494GCNDepth
3DKITTI Eigen split unsupervisedRMSE log0.181GCNDepth
3DKITTI Eigen split unsupervisedSq Rel0.72GCNDepth
3DKITTI Eigen split unsupervisedabsolute relative error0.104GCNDepth

Related Papers

AutoPartGen: Autogressive 3D Part Generation and Discovery2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16BRUM: Robust 3D Vehicle Reconstruction from 360 Sparse Images2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16MonoMVSNet: Monocular Priors Guided Multi-View Stereo Network2025-07-15