TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Simple Baseline for Efficient Hand Mesh Reconstruction

A Simple Baseline for Efficient Hand Mesh Reconstruction

Zhishan Zhou, Shihao. zhou, Zhi Lv, Minqiang Zou, Yao Tang, Jiajun Liang

2024-03-04CVPR 2024 13D Hand Pose EstimationPose EstimationGesture RecognitionHand Pose Estimation
PaperPDFCode

Abstract

3D hand pose estimation has found broad application in areas such as gesture recognition and human-machine interaction tasks. As performance improves, the complexity of the systems also increases, which can limit the comparative analysis and practical implementation of these methods. In this paper, we propose a simple yet effective baseline that not only surpasses state-of-the-art (SOTA) methods but also demonstrates computational efficiency. To establish this baseline, we abstract existing work into two components: a token generator and a mesh regressor, and then examine their core structures. A core structure, in this context, is one that fulfills intrinsic functions, brings about significant improvements, and achieves excellent performance without unnecessary complexities. Our proposed approach is decoupled from any modifications to the backbone, making it adaptable to any modern models. Our method outperforms existing solutions, achieving state-of-the-art (SOTA) results across multiple datasets. On the FreiHAND dataset, our approach produced a PA-MPJPE of 5.7mm and a PA-MPVPE of 6.0mm. Similarly, on the Dexycb dataset, we observed a PA-MPJPE of 5.5mm and a PA-MPVPE of 5.0mm. As for performance speed, our method reached up to 33 frames per second (fps) when using HRNet and up to 70 fps when employing FastViT-MA36

Results

TaskDatasetMetricValueModel
HandFreiHANDPA-F@15mm0.986Zhou et al.
HandFreiHANDPA-F@5mm0.772Zhou et al.
HandFreiHANDPA-MPJPE5.7Zhou et al.
HandFreiHANDPA-MPVPE6Zhou et al.
HandDexYCBAverage MPJPE (mm)12.4SimpleHand
HandDexYCBMPVPE12.1SimpleHand
HandDexYCBPA-MPVPE5.5SimpleHand
HandDexYCBProcrustes-Aligned MPJPE5.5SimpleHand
Pose EstimationFreiHANDPA-F@15mm0.986Zhou et al.
Pose EstimationFreiHANDPA-F@5mm0.772Zhou et al.
Pose EstimationFreiHANDPA-MPJPE5.7Zhou et al.
Pose EstimationFreiHANDPA-MPVPE6Zhou et al.
Pose EstimationDexYCBAverage MPJPE (mm)12.4SimpleHand
Pose EstimationDexYCBMPVPE12.1SimpleHand
Pose EstimationDexYCBPA-MPVPE5.5SimpleHand
Pose EstimationDexYCBProcrustes-Aligned MPJPE5.5SimpleHand
Hand Pose EstimationFreiHANDPA-F@15mm0.986Zhou et al.
Hand Pose EstimationFreiHANDPA-F@5mm0.772Zhou et al.
Hand Pose EstimationFreiHANDPA-MPJPE5.7Zhou et al.
Hand Pose EstimationFreiHANDPA-MPVPE6Zhou et al.
Hand Pose EstimationDexYCBAverage MPJPE (mm)12.4SimpleHand
Hand Pose EstimationDexYCBMPVPE12.1SimpleHand
Hand Pose EstimationDexYCBPA-MPVPE5.5SimpleHand
Hand Pose EstimationDexYCBProcrustes-Aligned MPJPE5.5SimpleHand
3DFreiHANDPA-F@15mm0.986Zhou et al.
3DFreiHANDPA-F@5mm0.772Zhou et al.
3DFreiHANDPA-MPJPE5.7Zhou et al.
3DFreiHANDPA-MPVPE6Zhou et al.
3DDexYCBAverage MPJPE (mm)12.4SimpleHand
3DDexYCBMPVPE12.1SimpleHand
3DDexYCBPA-MPVPE5.5SimpleHand
3DDexYCBProcrustes-Aligned MPJPE5.5SimpleHand
3D Hand Pose EstimationFreiHANDPA-F@15mm0.986Zhou et al.
3D Hand Pose EstimationFreiHANDPA-F@5mm0.772Zhou et al.
3D Hand Pose EstimationFreiHANDPA-MPJPE5.7Zhou et al.
3D Hand Pose EstimationFreiHANDPA-MPVPE6Zhou et al.
3D Hand Pose EstimationDexYCBAverage MPJPE (mm)12.4SimpleHand
3D Hand Pose EstimationDexYCBMPVPE12.1SimpleHand
3D Hand Pose EstimationDexYCBPA-MPVPE5.5SimpleHand
3D Hand Pose EstimationDexYCBProcrustes-Aligned MPJPE5.5SimpleHand
1 Image, 2*2 StitchiFreiHANDPA-F@15mm0.986Zhou et al.
1 Image, 2*2 StitchiFreiHANDPA-F@5mm0.772Zhou et al.
1 Image, 2*2 StitchiFreiHANDPA-MPJPE5.7Zhou et al.
1 Image, 2*2 StitchiFreiHANDPA-MPVPE6Zhou et al.
1 Image, 2*2 StitchiDexYCBAverage MPJPE (mm)12.4SimpleHand
1 Image, 2*2 StitchiDexYCBMPVPE12.1SimpleHand
1 Image, 2*2 StitchiDexYCBPA-MPVPE5.5SimpleHand
1 Image, 2*2 StitchiDexYCBProcrustes-Aligned MPJPE5.5SimpleHand

Related Papers

Efficient Deployment of Spiking Neural Networks on SpiNNaker2 for DVS Gesture Recognition Using Neuromorphic Intermediate Representation2025-09-04$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16