TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/(Fusionformer):Exploiting the Joint Motion Synergy with Fu...

(Fusionformer):Exploiting the Joint Motion Synergy with Fusion Network Based On Transformer for 3D Human Pose Estimation

Xinwei Yu, Xiaohua Zhang

2022-10-083D Human Pose EstimationPose Estimation
PaperPDF

Abstract

For the current 3D human pose estimation task, a group of methods mainly learn the rules of 2D-3D projection from spatial and temporal correlation. However, earlier methods model the global features of the entire body joint in the time domain, but ignore the motion trajectory of individual joint. The recent work [29] considers that there are differences in motion between different joints and deals with the temporal relationship of each joint separately. However, we found that different joints show the same movement trends under some specific actions. Therefore, our proposed Fusionformer method introduces a self-trajectory module and a mutual-trajectory module based on the spatio-temporal module .After that, the global spatio-temporal features and local joint trajectory features are fused through a linear network in a parallel manner. To eliminate the influence of bad 2D poses on 3D projections, finally we also introduce a pose refinement network to balance the consistency of 3D projections. In addition, we evaluate the proposed method on two benchmark datasets (Human3.6M, MPI-INF-3DHP). Comparing our method with the baseline method poseformer, the results show an improvement of 2.4% MPJPE and 4.3% P-MPJPE on the Human3.6M dataset, respectively.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationMPI-INF-3DHPAUC70Fusionformer (f=9)
3D Human Pose EstimationMPI-INF-3DHPMPJPE28.2Fusionformer (f=9)
3D Human Pose EstimationMPI-INF-3DHPPCK97.9Fusionformer (f=9)
Pose EstimationMPI-INF-3DHPAUC70Fusionformer (f=9)
Pose EstimationMPI-INF-3DHPMPJPE28.2Fusionformer (f=9)
Pose EstimationMPI-INF-3DHPPCK97.9Fusionformer (f=9)
3DMPI-INF-3DHPAUC70Fusionformer (f=9)
3DMPI-INF-3DHPMPJPE28.2Fusionformer (f=9)
3DMPI-INF-3DHPPCK97.9Fusionformer (f=9)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC70Fusionformer (f=9)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE28.2Fusionformer (f=9)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK97.9Fusionformer (f=9)

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16