TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Collaborative Learning for Hand and Object Reconstruction ...

Collaborative Learning for Hand and Object Reconstruction with Attention-guided Graph Convolution

Tze Ho Elden Tse, Kwang In Kim, Ales Leonardis, Hyung Jin Chang

2022-04-27CVPR 2022 13D Hand Pose EstimationObject Reconstructionhand-object posePose Estimation3D Pose Estimation
PaperPDF

Abstract

Estimating the pose and shape of hands and objects under interaction finds numerous applications including augmented and virtual reality. Existing approaches for hand and object reconstruction require explicitly defined physical constraints and known objects, which limits its application domains. Our algorithm is agnostic to object models, and it learns the physical rules governing hand-object interaction. This requires automatically inferring the shapes and physical interaction of hands and (potentially unknown) objects. We seek to approach this challenging problem by proposing a collaborative learning strategy where two-branches of deep networks are learning from each other. Specifically, we transfer hand mesh information to the object branch and vice versa for the hand branch. The resulting optimisation (training) problem can be unstable, and we address this via two strategies: (i) attention-guided graph convolution which helps identify and focus on mutual occlusion and (ii) unsupervised associative loss which facilitates the transfer of information between the branches. Experiments using four widely-used benchmarks show that our framework achieves beyond state-of-the-art accuracy in 3D pose estimation, as well as recovers dense 3D hand and object shapes. Each technical component above contributes meaningfully in the ablation study.

Results

TaskDatasetMetricValueModel
HandHO-3D v2F@15mm0.943Tse et al.
HandHO-3D v2F@5mm0.485Tse et al.
HandHO-3D v2PA-MPVPE10.9Tse et al.
HandDexYCBAverage MPJPE (mm)15.3CLAGC
Pose EstimationDexYCBAverage MPJPE (mm)15.3CLAGC
Pose EstimationHO-3D v2F@15mm0.943Tse et al.
Pose EstimationHO-3D v2F@5mm0.485Tse et al.
Pose EstimationHO-3D v2PA-MPVPE10.9Tse et al.
Hand Pose EstimationHO-3D v2F@15mm0.943Tse et al.
Hand Pose EstimationHO-3D v2F@5mm0.485Tse et al.
Hand Pose EstimationHO-3D v2PA-MPVPE10.9Tse et al.
Hand Pose EstimationDexYCBAverage MPJPE (mm)15.3CLAGC
3DDexYCBAverage MPJPE (mm)15.3CLAGC
3DHO-3D v2F@15mm0.943Tse et al.
3DHO-3D v2F@5mm0.485Tse et al.
3DHO-3D v2PA-MPVPE10.9Tse et al.
3D Hand Pose EstimationHO-3D v2F@15mm0.943Tse et al.
3D Hand Pose EstimationHO-3D v2F@5mm0.485Tse et al.
3D Hand Pose EstimationHO-3D v2PA-MPVPE10.9Tse et al.
3D Hand Pose EstimationDexYCBAverage MPJPE (mm)15.3CLAGC
6D Pose EstimationDexYCBAverage MPJPE (mm)15.3CLAGC
1 Image, 2*2 StitchiDexYCBAverage MPJPE (mm)15.3CLAGC
1 Image, 2*2 StitchiHO-3D v2F@15mm0.943Tse et al.
1 Image, 2*2 StitchiHO-3D v2F@5mm0.485Tse et al.
1 Image, 2*2 StitchiHO-3D v2PA-MPVPE10.9Tse et al.

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16