TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Interacting Hand-Object Pose Estimation via Dense Mutual A...

Interacting Hand-Object Pose Estimation via Dense Mutual Attention

Rong Wang, Wei Mao, Hongdong Li

2022-11-163D Hand Pose Estimationhand-object posePose Estimation
PaperPDFCode(official)

Abstract

3D hand-object pose estimation is the key to the success of many computer vision applications. The main focus of this task is to effectively model the interaction between the hand and an object. To this end, existing works either rely on interaction constraints in a computationally-expensive iterative optimization, or consider only a sparse correlation between sampled hand and object keypoints. In contrast, we propose a novel dense mutual attention mechanism that is able to model fine-grained dependencies between the hand and the object. Specifically, we first construct the hand and object graphs according to their mesh structures. For each hand node, we aggregate features from every object node by the learned attention and vice versa for each object node. Thanks to such dense mutual attention, our method is able to produce physically plausible poses with high quality and real-time inference speed. Extensive quantitative and qualitative experiments on large benchmark datasets show that our method outperforms state-of-the-art methods. The code is available at https://github.com/rongakowang/DenseMutualAttention.git.

Results

TaskDatasetMetricValueModel
HandHO-3D v2PA-MPJPE (mm)10.1DMA
HandHO-3D v2ADD-S20.8DMA
HandHO-3D v2Average MPJPE (mm)22.2DMA
HandHO-3D v2OME45.5DMA
HandHO-3D v2PA-MPJPE10.1DMA
HandHO-3D v2ST-MPJPE23.8DMA
HandDexYCBADD-S15.9DMA
HandDexYCBAverage MPJPE (mm)12.7DMA
HandDexYCBMCE32.6DMA
HandDexYCBOCE27.3DMA
HandDexYCBProcrustes-Aligned MPJPE6.86DMA
Pose EstimationHO-3D v2ADD-S20.8DMA
Pose EstimationHO-3D v2Average MPJPE (mm)22.2DMA
Pose EstimationHO-3D v2OME45.5DMA
Pose EstimationHO-3D v2PA-MPJPE10.1DMA
Pose EstimationHO-3D v2ST-MPJPE23.8DMA
Pose EstimationDexYCBADD-S15.9DMA
Pose EstimationDexYCBAverage MPJPE (mm)12.7DMA
Pose EstimationDexYCBMCE32.6DMA
Pose EstimationDexYCBOCE27.3DMA
Pose EstimationDexYCBProcrustes-Aligned MPJPE6.86DMA
Pose EstimationHO-3D v2PA-MPJPE (mm)10.1DMA
Hand Pose EstimationHO-3D v2PA-MPJPE (mm)10.1DMA
Hand Pose EstimationHO-3D v2ADD-S20.8DMA
Hand Pose EstimationHO-3D v2Average MPJPE (mm)22.2DMA
Hand Pose EstimationHO-3D v2OME45.5DMA
Hand Pose EstimationHO-3D v2PA-MPJPE10.1DMA
Hand Pose EstimationHO-3D v2ST-MPJPE23.8DMA
Hand Pose EstimationDexYCBADD-S15.9DMA
Hand Pose EstimationDexYCBAverage MPJPE (mm)12.7DMA
Hand Pose EstimationDexYCBMCE32.6DMA
Hand Pose EstimationDexYCBOCE27.3DMA
Hand Pose EstimationDexYCBProcrustes-Aligned MPJPE6.86DMA
3DHO-3D v2ADD-S20.8DMA
3DHO-3D v2Average MPJPE (mm)22.2DMA
3DHO-3D v2OME45.5DMA
3DHO-3D v2PA-MPJPE10.1DMA
3DHO-3D v2ST-MPJPE23.8DMA
3DDexYCBADD-S15.9DMA
3DDexYCBAverage MPJPE (mm)12.7DMA
3DDexYCBMCE32.6DMA
3DDexYCBOCE27.3DMA
3DDexYCBProcrustes-Aligned MPJPE6.86DMA
3DHO-3D v2PA-MPJPE (mm)10.1DMA
3D Hand Pose EstimationHO-3D v2PA-MPJPE (mm)10.1DMA
3D Hand Pose EstimationHO-3D v2ADD-S20.8DMA
3D Hand Pose EstimationHO-3D v2Average MPJPE (mm)22.2DMA
3D Hand Pose EstimationHO-3D v2OME45.5DMA
3D Hand Pose EstimationHO-3D v2PA-MPJPE10.1DMA
3D Hand Pose EstimationHO-3D v2ST-MPJPE23.8DMA
3D Hand Pose EstimationDexYCBADD-S15.9DMA
3D Hand Pose EstimationDexYCBAverage MPJPE (mm)12.7DMA
3D Hand Pose EstimationDexYCBMCE32.6DMA
3D Hand Pose EstimationDexYCBOCE27.3DMA
3D Hand Pose EstimationDexYCBProcrustes-Aligned MPJPE6.86DMA
6D Pose EstimationHO-3D v2ADD-S20.8DMA
6D Pose EstimationHO-3D v2Average MPJPE (mm)22.2DMA
6D Pose EstimationHO-3D v2OME45.5DMA
6D Pose EstimationHO-3D v2PA-MPJPE10.1DMA
6D Pose EstimationHO-3D v2ST-MPJPE23.8DMA
6D Pose EstimationDexYCBADD-S15.9DMA
6D Pose EstimationDexYCBAverage MPJPE (mm)12.7DMA
6D Pose EstimationDexYCBMCE32.6DMA
6D Pose EstimationDexYCBOCE27.3DMA
6D Pose EstimationDexYCBProcrustes-Aligned MPJPE6.86DMA
1 Image, 2*2 StitchiHO-3D v2ADD-S20.8DMA
1 Image, 2*2 StitchiHO-3D v2Average MPJPE (mm)22.2DMA
1 Image, 2*2 StitchiHO-3D v2OME45.5DMA
1 Image, 2*2 StitchiHO-3D v2PA-MPJPE10.1DMA
1 Image, 2*2 StitchiHO-3D v2ST-MPJPE23.8DMA
1 Image, 2*2 StitchiDexYCBADD-S15.9DMA
1 Image, 2*2 StitchiDexYCBAverage MPJPE (mm)12.7DMA
1 Image, 2*2 StitchiDexYCBMCE32.6DMA
1 Image, 2*2 StitchiDexYCBOCE27.3DMA
1 Image, 2*2 StitchiDexYCBProcrustes-Aligned MPJPE6.86DMA
1 Image, 2*2 StitchiHO-3D v2PA-MPJPE (mm)10.1DMA

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16