TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TransPose: Keypoint Localization via Transformer

TransPose: Keypoint Localization via Transformer

Sen yang, Zhibin Quan, Mu Nie, Wankou Yang

2020-12-28ICCV 2021 10Pose EstimationMulti-Person Pose EstimationKeypoint Detection
PaperPDFCode(official)

Abstract

While CNN-based models have made remarkable progress on human pose estimation, what spatial dependencies they capture to localize keypoints remains unclear. In this work, we propose a model called \textbf{TransPose}, which introduces Transformer for human pose estimation. The attention layers built in Transformer enable our model to capture long-range relationships efficiently and also can reveal what dependencies the predicted keypoints rely on. To predict keypoint heatmaps, the last attention layer acts as an aggregator, which collects contributions from image clues and forms maximum positions of keypoints. Such a heatmap-based localization approach via Transformer conforms to the principle of Activation Maximization~\cite{erhan2009visualizing}. And the revealed dependencies are image-specific and fine-grained, which also can provide evidence of how the model handles special cases, e.g., occlusion. The experiments show that TransPose achieves 75.8 AP and 75.0 AP on COCO validation and test-dev sets, while being more lightweight and faster than mainstream CNN architectures. The TransPose model also transfers very well on MPII benchmark, achieving superior performance on the test set when fine-tuned with small training costs. Code and pre-trained models are publicly available\footnote{\url{https://github.com/yangsenius/TransPose}}.

Results

TaskDatasetMetricValueModel
Pose EstimationOCHumanValidation AP62.3TransPose-H
Pose EstimationCOCO test-devAP75TransPose-H-A6
Pose EstimationCOCO test-devAP5092.2TransPose-H-A6
Pose EstimationCOCO test-devAP7582.3TransPose-H-A6
Pose EstimationCOCO test-devAPL81.1TransPose-H-A6
Pose EstimationCOCO test-devAPM71.3TransPose-H-A6
Pose EstimationMPII Human PosePCKh-0.593.5TransPose
Pose EstimationCOCO (Common Objects in Context)Test AP75TransPose(256x192)
Pose EstimationCOCO (Common Objects in Context)Validation AP75.8TransPose(256x192)
Pose EstimationCrowdPoseAP Easy79.5TransPose-H
Pose EstimationCrowdPoseAP Hard62.2TransPose-H
Pose EstimationCrowdPoseAP Medium72.9TransPose-H
Pose EstimationCrowdPosemAP @0.5:0.9571.8TransPose-H
Pose EstimationOCHumanAP5082.7TransPose-H
Pose EstimationOCHumanAP7567.1TransPose-H
3DOCHumanValidation AP62.3TransPose-H
3DCOCO test-devAP75TransPose-H-A6
3DCOCO test-devAP5092.2TransPose-H-A6
3DCOCO test-devAP7582.3TransPose-H-A6
3DCOCO test-devAPL81.1TransPose-H-A6
3DCOCO test-devAPM71.3TransPose-H-A6
3DMPII Human PosePCKh-0.593.5TransPose
3DCOCO (Common Objects in Context)Test AP75TransPose(256x192)
3DCOCO (Common Objects in Context)Validation AP75.8TransPose(256x192)
3DCrowdPoseAP Easy79.5TransPose-H
3DCrowdPoseAP Hard62.2TransPose-H
3DCrowdPoseAP Medium72.9TransPose-H
3DCrowdPosemAP @0.5:0.9571.8TransPose-H
3DOCHumanAP5082.7TransPose-H
3DOCHumanAP7567.1TransPose-H
Multi-Person Pose EstimationCrowdPoseAP Easy79.5TransPose-H
Multi-Person Pose EstimationCrowdPoseAP Hard62.2TransPose-H
Multi-Person Pose EstimationCrowdPoseAP Medium72.9TransPose-H
Multi-Person Pose EstimationCrowdPosemAP @0.5:0.9571.8TransPose-H
Multi-Person Pose EstimationOCHumanAP5082.7TransPose-H
Multi-Person Pose EstimationOCHumanAP7567.1TransPose-H
1 Image, 2*2 StitchiOCHumanValidation AP62.3TransPose-H
1 Image, 2*2 StitchiCOCO test-devAP75TransPose-H-A6
1 Image, 2*2 StitchiCOCO test-devAP5092.2TransPose-H-A6
1 Image, 2*2 StitchiCOCO test-devAP7582.3TransPose-H-A6
1 Image, 2*2 StitchiCOCO test-devAPL81.1TransPose-H-A6
1 Image, 2*2 StitchiCOCO test-devAPM71.3TransPose-H-A6
1 Image, 2*2 StitchiMPII Human PosePCKh-0.593.5TransPose
1 Image, 2*2 StitchiCOCO (Common Objects in Context)Test AP75TransPose(256x192)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)Validation AP75.8TransPose(256x192)
1 Image, 2*2 StitchiCrowdPoseAP Easy79.5TransPose-H
1 Image, 2*2 StitchiCrowdPoseAP Hard62.2TransPose-H
1 Image, 2*2 StitchiCrowdPoseAP Medium72.9TransPose-H
1 Image, 2*2 StitchiCrowdPosemAP @0.5:0.9571.8TransPose-H
1 Image, 2*2 StitchiOCHumanAP5082.7TransPose-H
1 Image, 2*2 StitchiOCHumanAP7567.1TransPose-H

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16