TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Decoupled Iterative Refinement Framework for Interacting H...

Decoupled Iterative Refinement Framework for Interacting Hands Reconstruction from a Single RGB Image

Pengfei Ren, Chao Wen, Xiaozheng Zheng, Zhou Xue, Haifeng Sun, Qi Qi, Jingyu Wang, Jianxin Liao

2023-02-05ICCV 2023 13D Hand Pose Estimation3D Interacting Hand Pose Estimation
PaperPDFCode(official)

Abstract

Reconstructing interacting hands from a single RGB image is a very challenging task. On the one hand, severe mutual occlusion and similar local appearance between two hands confuse the extraction of visual features, resulting in the misalignment of estimated hand meshes and the image. On the other hand, there are complex spatial relationship between interacting hands, which significantly increases the solution space of hand poses and increases the difficulty of network learning. In this paper, we propose a decoupled iterative refinement framework to achieve pixel-alignment hand reconstruction while efficiently modeling the spatial relationship between hands. Specifically, we define two feature spaces with different characteristics, namely 2D visual feature space and 3D joint feature space. First, we obtain joint-wise features from the visual feature map and utilize a graph convolution network and a transformer to perform intra- and inter-hand information interaction in the 3D joint feature space, respectively. Then, we project the joint features with global information back into the 2D visual feature space in an obfuscation-free manner and utilize the 2D convolution for pixel-wise enhancement. By performing multiple alternate enhancements in the two feature spaces, our method can achieve an accurate and robust reconstruction of interacting hands. Our method outperforms all existing two-hand reconstruction methods by a large margin on the InterHand2.6M dataset.

Results

TaskDatasetMetricValueModel
3D Interacting Hand Pose EstimationInterHand2.6MMPJPE Test7.51DIR
3D Interacting Hand Pose EstimationInterHand2.6MMPVPE Test7.72DIR
3D Interacting Hand Pose EstimationInterHand2.6MMRRPE Test28.98DIR

Related Papers

ExtPose: Robust and Coherent Pose Estimation by Extending ViTs2025-06-18Monocular 3D Hand Pose Estimation with Implicit Camera Alignment2025-06-10VM-BHINet:Vision Mamba Bimanual Hand Interaction Network for 3D Interacting Hand Mesh Recovery From a Single RGB Image2025-04-20OccRobNet : Occlusion Robust Network for Accurate 3D Interacting Hand-Object Pose Estimation2025-03-27Analyzing the Synthetic-to-Real Domain Gap in 3D Hand Pose Estimation2025-03-25SiMHand: Mining Similar Hands for Large-Scale 3D Hand Pose Pre-training2025-02-21HaWoR: World-Space Hand Motion Reconstruction from Egocentric Videos2025-01-06BIGS: Bimanual Category-agnostic Interaction Reconstruction from Monocular Videos via 3D Gaussian Splatting2025-01-01