TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Leveraging Photometric Consistency over Time for Sparsely ...

Leveraging Photometric Consistency over Time for Sparsely Supervised Hand-Object Reconstruction

Yana Hasson, Bugra Tekin, Federica Bogo, Ivan Laptev, Marc Pollefeys, Cordelia Schmid

2020-04-28CVPR 2020 6Optical Flow EstimationObject Reconstructionhand-object posePose Estimation
PaperPDF

Abstract

Modeling hand-object manipulations is essential for understanding how humans interact with their environment. While of practical importance, estimating the pose of hands and objects during interactions is challenging due to the large mutual occlusions that occur during manipulation. Recent efforts have been directed towards fully-supervised methods that require large amounts of labeled training samples. Collecting 3D ground-truth data for hand-object interactions, however, is costly, tedious, and error-prone. To overcome this challenge we present a method to leverage photometric consistency across time when annotations are only available for a sparse subset of frames in a video. Our model is trained end-to-end on color images to jointly reconstruct hands and objects in 3D by inferring their poses. Given our estimated reconstructions, we differentiably render the optical flow between pairs of adjacent images and use it within the network to warp one frame to another. We then apply a self-supervised photometric loss that relies on the visual consistency between nearby images. We achieve state-of-the-art results on 3D hand-object reconstruction benchmarks and demonstrate that our approach allows us to improve the pose estimation accuracy by leveraging information from neighboring frames in low-data regimes.

Results

TaskDatasetMetricValueModel
HandHO-3D v2ADD-S22PCTHO
HandHO-3D v2OME67PCTHO
HandHO-3D v2PA-MPJPE11.4PCTHO
HandHO-3D v2ST-MPJPE36.9PCTHO
Pose EstimationHO-3D v2ADD-S22PCTHO
Pose EstimationHO-3D v2OME67PCTHO
Pose EstimationHO-3D v2PA-MPJPE11.4PCTHO
Pose EstimationHO-3D v2ST-MPJPE36.9PCTHO
Hand Pose EstimationHO-3D v2ADD-S22PCTHO
Hand Pose EstimationHO-3D v2OME67PCTHO
Hand Pose EstimationHO-3D v2PA-MPJPE11.4PCTHO
Hand Pose EstimationHO-3D v2ST-MPJPE36.9PCTHO
3DHO-3D v2ADD-S22PCTHO
3DHO-3D v2OME67PCTHO
3DHO-3D v2PA-MPJPE11.4PCTHO
3DHO-3D v2ST-MPJPE36.9PCTHO
3D Hand Pose EstimationHO-3D v2ADD-S22PCTHO
3D Hand Pose EstimationHO-3D v2OME67PCTHO
3D Hand Pose EstimationHO-3D v2PA-MPJPE11.4PCTHO
3D Hand Pose EstimationHO-3D v2ST-MPJPE36.9PCTHO
6D Pose EstimationHO-3D v2ADD-S22PCTHO
6D Pose EstimationHO-3D v2OME67PCTHO
6D Pose EstimationHO-3D v2PA-MPJPE11.4PCTHO
6D Pose EstimationHO-3D v2ST-MPJPE36.9PCTHO
1 Image, 2*2 StitchiHO-3D v2ADD-S22PCTHO
1 Image, 2*2 StitchiHO-3D v2OME67PCTHO
1 Image, 2*2 StitchiHO-3D v2PA-MPJPE11.4PCTHO
1 Image, 2*2 StitchiHO-3D v2ST-MPJPE36.9PCTHO

Related Papers

Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16