TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TriHorn-Net: A Model for Accurate Depth-Based 3D Hand Pose...

TriHorn-Net: A Model for Accurate Depth-Based 3D Hand Pose Estimation

Mohammad Rezaei, Razieh Rastgoo, Vassilis Athitsos

2022-06-143D Hand Pose EstimationData AugmentationPose EstimationDepth EstimationHand Pose Estimation
PaperPDFCode(official)

Abstract

3D hand pose estimation methods have made significant progress recently. However, the estimation accuracy is often far from sufficient for specific real-world applications, and thus there is significant room for improvement. This paper proposes TriHorn-Net, a novel model that uses specific innovations to improve hand pose estimation accuracy on depth images. The first innovation is the decomposition of the 3D hand pose estimation into the estimation of 2D joint locations in the depth image space (UV), and the estimation of their corresponding depths aided by two complementary attention maps. This decomposition prevents depth estimation, which is a more difficult task, from interfering with the UV estimations at both the prediction and feature levels. The second innovation is PixDropout, which is, to the best of our knowledge, the first appearance-based data augmentation method for hand depth images. Experimental results demonstrate that the proposed model outperforms the state-of-the-art methods on three public benchmark datasets. Our implementation is available at https://github.com/mrezaei92/TriHorn-Net.

Results

TaskDatasetMetricValueModel
HandMSRA HandsAverage 3D Error7.13TriHorn-Net
HandICVL HandsAverage 3D Error5.73TriHorn-Net
HandNYU HandsAverage 3D Error7.68TriHorn-Net
Pose EstimationMSRA HandsAverage 3D Error7.13TriHorn-Net
Pose EstimationICVL HandsAverage 3D Error5.73TriHorn-Net
Pose EstimationNYU HandsAverage 3D Error7.68TriHorn-Net
Hand Pose EstimationMSRA HandsAverage 3D Error7.13TriHorn-Net
Hand Pose EstimationICVL HandsAverage 3D Error5.73TriHorn-Net
Hand Pose EstimationNYU HandsAverage 3D Error7.68TriHorn-Net
3DMSRA HandsAverage 3D Error7.13TriHorn-Net
3DICVL HandsAverage 3D Error5.73TriHorn-Net
3DNYU HandsAverage 3D Error7.68TriHorn-Net
1 Image, 2*2 StitchiMSRA HandsAverage 3D Error7.13TriHorn-Net
1 Image, 2*2 StitchiICVL HandsAverage 3D Error5.73TriHorn-Net
1 Image, 2*2 StitchiNYU HandsAverage 3D Error7.68TriHorn-Net

Related Papers

Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17$S^2M^2$: Scalable Stereo Matching Model for Reliable Depth Estimation2025-07-17