TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Model-based 3D Hand Reconstruction via Self-Supervised Lea...

Model-based 3D Hand Reconstruction via Self-Supervised Learning

Yujin Chen, Zhigang Tu, Di Kang, Linchao Bao, Ying Zhang, Xuefei Zhe, Ruizhi Chen, Junsong Yuan

2021-03-22CVPR 2021 13D Hand Pose EstimationSelf-Supervised Learning
PaperPDFCode(official)

Abstract

Reconstructing a 3D hand from a single-view RGB image is challenging due to various hand configurations and depth ambiguity. To reliably reconstruct a 3D hand from a monocular image, most state-of-the-art methods heavily rely on 3D annotations at the training stage, but obtaining 3D annotations is expensive. To alleviate reliance on labeled training data, we propose S2HAND, a self-supervised 3D hand reconstruction network that can jointly estimate pose, shape, texture, and the camera viewpoint. Specifically, we obtain geometric cues from the input image through easily accessible 2D detected keypoints. To learn an accurate hand reconstruction model from these noisy geometric cues, we utilize the consistency between 2D and 3D representations and propose a set of novel losses to rationalize outputs of the neural network. For the first time, we demonstrate the feasibility of training an accurate 3D hand reconstruction network without relying on manual annotations. Our experiments show that the proposed method achieves comparable performance with recent fully-supervised methods while using fewer supervision data.

Results

TaskDatasetMetricValueModel
HandHO-3D v3AUC_J0.769S2HAND
HandHO-3D v3AUC_V0.778S2HAND
HandHO-3D v3F@15mm0.932S2HAND
HandHO-3D v3F@5mm0.448S2HAND
HandHO-3D v3PA-MPJPE11.5S2HAND
HandHO-3D v3PA-MPVPE11.1S2HAND
HandHO-3D v2AUC_J0.773S2Hand
HandHO-3D v2AUC_V0.777S2Hand
HandHO-3D v2F@15mm0.93S2Hand
HandHO-3D v2F@5mm0.45S2Hand
HandHO-3D v2PA-MPJPE (mm)11.4S2Hand
HandHO-3D v2PA-MPVPE11.2S2Hand
Pose EstimationHO-3D v3AUC_J0.769S2HAND
Pose EstimationHO-3D v3AUC_V0.778S2HAND
Pose EstimationHO-3D v3F@15mm0.932S2HAND
Pose EstimationHO-3D v3F@5mm0.448S2HAND
Pose EstimationHO-3D v3PA-MPJPE11.5S2HAND
Pose EstimationHO-3D v3PA-MPVPE11.1S2HAND
Pose EstimationHO-3D v2AUC_J0.773S2Hand
Pose EstimationHO-3D v2AUC_V0.777S2Hand
Pose EstimationHO-3D v2F@15mm0.93S2Hand
Pose EstimationHO-3D v2F@5mm0.45S2Hand
Pose EstimationHO-3D v2PA-MPJPE (mm)11.4S2Hand
Pose EstimationHO-3D v2PA-MPVPE11.2S2Hand
Hand Pose EstimationHO-3D v3AUC_J0.769S2HAND
Hand Pose EstimationHO-3D v3AUC_V0.778S2HAND
Hand Pose EstimationHO-3D v3F@15mm0.932S2HAND
Hand Pose EstimationHO-3D v3F@5mm0.448S2HAND
Hand Pose EstimationHO-3D v3PA-MPJPE11.5S2HAND
Hand Pose EstimationHO-3D v3PA-MPVPE11.1S2HAND
Hand Pose EstimationHO-3D v2AUC_J0.773S2Hand
Hand Pose EstimationHO-3D v2AUC_V0.777S2Hand
Hand Pose EstimationHO-3D v2F@15mm0.93S2Hand
Hand Pose EstimationHO-3D v2F@5mm0.45S2Hand
Hand Pose EstimationHO-3D v2PA-MPJPE (mm)11.4S2Hand
Hand Pose EstimationHO-3D v2PA-MPVPE11.2S2Hand
3DHO-3D v3AUC_J0.769S2HAND
3DHO-3D v3AUC_V0.778S2HAND
3DHO-3D v3F@15mm0.932S2HAND
3DHO-3D v3F@5mm0.448S2HAND
3DHO-3D v3PA-MPJPE11.5S2HAND
3DHO-3D v3PA-MPVPE11.1S2HAND
3DHO-3D v2AUC_J0.773S2Hand
3DHO-3D v2AUC_V0.777S2Hand
3DHO-3D v2F@15mm0.93S2Hand
3DHO-3D v2F@5mm0.45S2Hand
3DHO-3D v2PA-MPJPE (mm)11.4S2Hand
3DHO-3D v2PA-MPVPE11.2S2Hand
3D Hand Pose EstimationHO-3D v3AUC_J0.769S2HAND
3D Hand Pose EstimationHO-3D v3AUC_V0.778S2HAND
3D Hand Pose EstimationHO-3D v3F@15mm0.932S2HAND
3D Hand Pose EstimationHO-3D v3F@5mm0.448S2HAND
3D Hand Pose EstimationHO-3D v3PA-MPJPE11.5S2HAND
3D Hand Pose EstimationHO-3D v3PA-MPVPE11.1S2HAND
3D Hand Pose EstimationHO-3D v2AUC_J0.773S2Hand
3D Hand Pose EstimationHO-3D v2AUC_V0.777S2Hand
3D Hand Pose EstimationHO-3D v2F@15mm0.93S2Hand
3D Hand Pose EstimationHO-3D v2F@5mm0.45S2Hand
3D Hand Pose EstimationHO-3D v2PA-MPJPE (mm)11.4S2Hand
3D Hand Pose EstimationHO-3D v2PA-MPVPE11.2S2Hand
1 Image, 2*2 StitchiHO-3D v3AUC_J0.769S2HAND
1 Image, 2*2 StitchiHO-3D v3AUC_V0.778S2HAND
1 Image, 2*2 StitchiHO-3D v3F@15mm0.932S2HAND
1 Image, 2*2 StitchiHO-3D v3F@5mm0.448S2HAND
1 Image, 2*2 StitchiHO-3D v3PA-MPJPE11.5S2HAND
1 Image, 2*2 StitchiHO-3D v3PA-MPVPE11.1S2HAND
1 Image, 2*2 StitchiHO-3D v2AUC_J0.773S2Hand
1 Image, 2*2 StitchiHO-3D v2AUC_V0.777S2Hand
1 Image, 2*2 StitchiHO-3D v2F@15mm0.93S2Hand
1 Image, 2*2 StitchiHO-3D v2F@5mm0.45S2Hand
1 Image, 2*2 StitchiHO-3D v2PA-MPJPE (mm)11.4S2Hand
1 Image, 2*2 StitchiHO-3D v2PA-MPVPE11.2S2Hand

Related Papers

A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys2025-07-17Self-supervised Learning on Camera Trap Footage Yields a Strong Universal Face Embedder2025-07-14Speech Quality Assessment Model Based on Mixture of Experts: System-Level Performance Enhancement and Utterance-Level Challenge Analysis2025-07-08World4Drive: End-to-End Autonomous Driving via Intention-aware Physical Latent World Model2025-07-01ShapeEmbed: a self-supervised learning framework for 2D contour quantification2025-07-01RetFiner: A Vision-Language Refinement Scheme for Retinal Foundation Models2025-06-27Boosting Generative Adversarial Transferability with Self-supervised Vision Transformer Features2025-06-26Hybrid Deep Learning and Signal Processing for Arabic Dialect Recognition in Low-Resource Settings2025-06-26