TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MotionAGFormer: Enhancing 3D Human Pose Estimation with a ...

MotionAGFormer: Enhancing 3D Human Pose Estimation with a Transformer-GCNFormer Network

Soroush Mehraban, Vida Adeli, Babak Taati

2023-10-253D Human Pose EstimationMonocular 3D Human Pose EstimationPose EstimationClassification
PaperPDFCode(official)

Abstract

Recent transformer-based approaches have demonstrated excellent performance in 3D human pose estimation. However, they have a holistic view and by encoding global relationships between all the joints, they do not capture the local dependencies precisely. In this paper, we present a novel Attention-GCNFormer (AGFormer) block that divides the number of channels by using two parallel transformer and GCNFormer streams. Our proposed GCNFormer module exploits the local relationship between adjacent joints, outputting a new representation that is complementary to the transformer output. By fusing these two representation in an adaptive way, AGFormer exhibits the ability to better learn the underlying 3D structure. By stacking multiple AGFormer blocks, we propose MotionAGFormer in four different variants, which can be chosen based on the speed-accuracy trade-off. We evaluate our model on two popular benchmark datasets: Human3.6M and MPI-INF-3DHP. MotionAGFormer-B achieves state-of-the-art results, with P1 errors of 38.4mm and 16.2mm, respectively. Remarkably, it uses a quarter of the parameters and is three times more computationally efficient than the previous leading model on Human3.6M dataset. Code and models are available at https://github.com/TaatiTeam/MotionAGFormer.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationMPI-INF-3DHPAUC85.3MotionAGFormer-L (T=81)
3D Human Pose EstimationMPI-INF-3DHPMPJPE16.2MotionAGFormer-L (T=81)
3D Human Pose EstimationMPI-INF-3DHPPCK98.2MotionAGFormer-L (T=81)
3D Human Pose EstimationMPI-INF-3DHPAUC84.5MotionAGFormer-S (T=81)
3D Human Pose EstimationMPI-INF-3DHPMPJPE17.1MotionAGFormer-S (T=81)
3D Human Pose EstimationMPI-INF-3DHPPCK98.3MotionAGFormer-S (T=81)
3D Human Pose EstimationMPI-INF-3DHPAUC84.2MotionAGFormer-B (T=81)
3D Human Pose EstimationMPI-INF-3DHPMPJPE18.2MotionAGFormer-B (T=81)
3D Human Pose EstimationMPI-INF-3DHPPCK98.3MotionAGFormer-B (T=81)
3D Human Pose EstimationMPI-INF-3DHPAUC83.5MotionAGFormer-XS (T=27)
3D Human Pose EstimationMPI-INF-3DHPMPJPE19.2MotionAGFormer-XS (T=27)
3D Human Pose EstimationMPI-INF-3DHPPCK98.2MotionAGFormer-XS (T=27)
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-L
3D Human Pose EstimationHuman3.6MFrames Needed243MotionAGFormer-L
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-B
3D Human Pose EstimationHuman3.6MFrames Needed243MotionAGFormer-B
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)42.5MotionAGFormer-S
3D Human Pose EstimationHuman3.6MFrames Needed81MotionAGFormer-S
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)45.1MotionAGFormer-XS
3D Human Pose EstimationHuman3.6MFrames Needed27MotionAGFormer-XS
Pose EstimationMPI-INF-3DHPAUC85.3MotionAGFormer-L (T=81)
Pose EstimationMPI-INF-3DHPMPJPE16.2MotionAGFormer-L (T=81)
Pose EstimationMPI-INF-3DHPPCK98.2MotionAGFormer-L (T=81)
Pose EstimationMPI-INF-3DHPAUC84.5MotionAGFormer-S (T=81)
Pose EstimationMPI-INF-3DHPMPJPE17.1MotionAGFormer-S (T=81)
Pose EstimationMPI-INF-3DHPPCK98.3MotionAGFormer-S (T=81)
Pose EstimationMPI-INF-3DHPAUC84.2MotionAGFormer-B (T=81)
Pose EstimationMPI-INF-3DHPMPJPE18.2MotionAGFormer-B (T=81)
Pose EstimationMPI-INF-3DHPPCK98.3MotionAGFormer-B (T=81)
Pose EstimationMPI-INF-3DHPAUC83.5MotionAGFormer-XS (T=27)
Pose EstimationMPI-INF-3DHPMPJPE19.2MotionAGFormer-XS (T=27)
Pose EstimationMPI-INF-3DHPPCK98.2MotionAGFormer-XS (T=27)
Pose EstimationHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-L
Pose EstimationHuman3.6MFrames Needed243MotionAGFormer-L
Pose EstimationHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-B
Pose EstimationHuman3.6MFrames Needed243MotionAGFormer-B
Pose EstimationHuman3.6MAverage MPJPE (mm)42.5MotionAGFormer-S
Pose EstimationHuman3.6MFrames Needed81MotionAGFormer-S
Pose EstimationHuman3.6MAverage MPJPE (mm)45.1MotionAGFormer-XS
Pose EstimationHuman3.6MFrames Needed27MotionAGFormer-XS
3DMPI-INF-3DHPAUC85.3MotionAGFormer-L (T=81)
3DMPI-INF-3DHPMPJPE16.2MotionAGFormer-L (T=81)
3DMPI-INF-3DHPPCK98.2MotionAGFormer-L (T=81)
3DMPI-INF-3DHPAUC84.5MotionAGFormer-S (T=81)
3DMPI-INF-3DHPMPJPE17.1MotionAGFormer-S (T=81)
3DMPI-INF-3DHPPCK98.3MotionAGFormer-S (T=81)
3DMPI-INF-3DHPAUC84.2MotionAGFormer-B (T=81)
3DMPI-INF-3DHPMPJPE18.2MotionAGFormer-B (T=81)
3DMPI-INF-3DHPPCK98.3MotionAGFormer-B (T=81)
3DMPI-INF-3DHPAUC83.5MotionAGFormer-XS (T=27)
3DMPI-INF-3DHPMPJPE19.2MotionAGFormer-XS (T=27)
3DMPI-INF-3DHPPCK98.2MotionAGFormer-XS (T=27)
3DHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-L
3DHuman3.6MFrames Needed243MotionAGFormer-L
3DHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-B
3DHuman3.6MFrames Needed243MotionAGFormer-B
3DHuman3.6MAverage MPJPE (mm)42.5MotionAGFormer-S
3DHuman3.6MFrames Needed81MotionAGFormer-S
3DHuman3.6MAverage MPJPE (mm)45.1MotionAGFormer-XS
3DHuman3.6MFrames Needed27MotionAGFormer-XS
ClassificationFull-body Parkinson’s disease datasetF1-score (weighted)0.42MotionAGFormer
1 Image, 2*2 StitchiMPI-INF-3DHPAUC85.3MotionAGFormer-L (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE16.2MotionAGFormer-L (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK98.2MotionAGFormer-L (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC84.5MotionAGFormer-S (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE17.1MotionAGFormer-S (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK98.3MotionAGFormer-S (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC84.2MotionAGFormer-B (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE18.2MotionAGFormer-B (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK98.3MotionAGFormer-B (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC83.5MotionAGFormer-XS (T=27)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE19.2MotionAGFormer-XS (T=27)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK98.2MotionAGFormer-XS (T=27)
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-L
1 Image, 2*2 StitchiHuman3.6MFrames Needed243MotionAGFormer-L
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)38.4MotionAGFormer-B
1 Image, 2*2 StitchiHuman3.6MFrames Needed243MotionAGFormer-B
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)42.5MotionAGFormer-S
1 Image, 2*2 StitchiHuman3.6MFrames Needed81MotionAGFormer-S
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)45.1MotionAGFormer-XS
1 Image, 2*2 StitchiHuman3.6MFrames Needed27MotionAGFormer-XS

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16