TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Towards Accurate Multi-person Pose Estimation in the Wild

Towards Accurate Multi-person Pose Estimation in the Wild

George Papandreou, Tyler Zhu, Nori Kanazawa, Alexander Toshev, Jonathan Tompson, Chris Bregler, Kevin Murphy

2017-01-06CVPR 2017 7Human DetectionPose EstimationMulti-Person Pose EstimationKeypoint Detection
PaperPDF

Abstract

We propose a method for multi-person detection and 2-D pose estimation that achieves state-of-art results on the challenging COCO keypoints task. It is a simple, yet powerful, top-down approach consisting of two stages. In the first stage, we predict the location and scale of boxes which are likely to contain people; for this we use the Faster RCNN detector. In the second stage, we estimate the keypoints of the person potentially contained in each proposed bounding box. For each keypoint type we predict dense heatmaps and offsets using a fully convolutional ResNet. To combine these outputs we introduce a novel aggregation procedure to obtain highly localized keypoint predictions. We also use a novel form of keypoint-based Non-Maximum-Suppression (NMS), instead of the cruder box-level NMS, and a novel form of keypoint-based confidence score estimation, instead of box-level scoring. Trained on COCO data alone, our final system achieves average precision of 0.649 on the COCO test-dev set and the 0.643 test-standard sets, outperforming the winner of the 2016 COCO keypoints challenge and other recent state-of-art. Further, by using additional in-house labeled data we obtain an even higher average precision of 0.685 on the test-dev set and 0.673 on the test-standard set, more than 5% absolute improvement compared to the previous best performing method on the same dataset.

Results

TaskDatasetMetricValueModel
Pose EstimationCOCO test-devAP64.9G-RMI
Pose EstimationCOCO test-devAP5085.5G-RMI
Pose EstimationCOCO test-devAP7571.3G-RMI
Pose EstimationCOCO test-devAPL70G-RMI
Pose EstimationCOCO test-devAR69.7G-RMI
Pose EstimationCOCO test-devAP5085.5G-RMI
Pose EstimationCOCO test-devAP7571.3G-RMI
Pose EstimationCOCO test-devAPL70G-RMI
Pose EstimationCOCO test-devAPM62.3G-RMI
Pose EstimationCOCO test-devAR69.7G-RMI
Pose EstimationCOCO test-devAR5088.7G-RMI
Pose EstimationCOCO test-devAR7575.5G-RMI
Pose EstimationCOCO test-devARL77.1G-RMI
Pose EstimationCOCO test-devARM64.4G-RMI
Pose EstimationCOCO test-challengeAP69.1G-RMI*
Pose EstimationCOCO test-challengeAP5085.9G-RMI*
Pose EstimationCOCO test-challengeAP7575.2G-RMI*
Pose EstimationCOCO test-challengeAPL82.4G-RMI*
Pose EstimationCOCO test-challengeAR75.1G-RMI*
Pose EstimationCOCO test-challengeAR5090.7G-RMI*
Pose EstimationCOCO test-challengeAR7580.7G-RMI*
Pose EstimationCOCO test-challengeARL74.5G-RMI*
Pose EstimationCOCO test-challengeARM69.7G-RMI*
Pose EstimationCOCO test-devAP64.9G-RMI
Pose EstimationCOCO test-devAP5085.5G-RMI
Pose EstimationCOCO test-devAP7571.3G-RMI
Pose EstimationCOCO test-devAPL70G-RMI
Pose EstimationCOCO test-devAPM62.3G-RMI
Pose EstimationCOCO (Common Objects in Context)AP0.685G-RMI*
Pose EstimationCOCO (Common Objects in Context)AP0.649G-RMI
3DCOCO test-devAP64.9G-RMI
3DCOCO test-devAP5085.5G-RMI
3DCOCO test-devAP7571.3G-RMI
3DCOCO test-devAPL70G-RMI
3DCOCO test-devAR69.7G-RMI
3DCOCO test-devAP5085.5G-RMI
3DCOCO test-devAP7571.3G-RMI
3DCOCO test-devAPL70G-RMI
3DCOCO test-devAPM62.3G-RMI
3DCOCO test-devAR69.7G-RMI
3DCOCO test-devAR5088.7G-RMI
3DCOCO test-devAR7575.5G-RMI
3DCOCO test-devARL77.1G-RMI
3DCOCO test-devARM64.4G-RMI
3DCOCO test-challengeAP69.1G-RMI*
3DCOCO test-challengeAP5085.9G-RMI*
3DCOCO test-challengeAP7575.2G-RMI*
3DCOCO test-challengeAPL82.4G-RMI*
3DCOCO test-challengeAR75.1G-RMI*
3DCOCO test-challengeAR5090.7G-RMI*
3DCOCO test-challengeAR7580.7G-RMI*
3DCOCO test-challengeARL74.5G-RMI*
3DCOCO test-challengeARM69.7G-RMI*
3DCOCO test-devAP64.9G-RMI
3DCOCO test-devAP5085.5G-RMI
3DCOCO test-devAP7571.3G-RMI
3DCOCO test-devAPL70G-RMI
3DCOCO test-devAPM62.3G-RMI
3DCOCO (Common Objects in Context)AP0.685G-RMI*
3DCOCO (Common Objects in Context)AP0.649G-RMI
Multi-Person Pose EstimationCOCO test-devAP64.9G-RMI
Multi-Person Pose EstimationCOCO test-devAP5085.5G-RMI
Multi-Person Pose EstimationCOCO test-devAP7571.3G-RMI
Multi-Person Pose EstimationCOCO test-devAPL70G-RMI
Multi-Person Pose EstimationCOCO test-devAPM62.3G-RMI
Multi-Person Pose EstimationCOCO (Common Objects in Context)AP0.685G-RMI*
Multi-Person Pose EstimationCOCO (Common Objects in Context)AP0.649G-RMI
1 Image, 2*2 StitchiCOCO test-devAP64.9G-RMI
1 Image, 2*2 StitchiCOCO test-devAP5085.5G-RMI
1 Image, 2*2 StitchiCOCO test-devAP7571.3G-RMI
1 Image, 2*2 StitchiCOCO test-devAPL70G-RMI
1 Image, 2*2 StitchiCOCO test-devAR69.7G-RMI
1 Image, 2*2 StitchiCOCO test-devAP5085.5G-RMI
1 Image, 2*2 StitchiCOCO test-devAP7571.3G-RMI
1 Image, 2*2 StitchiCOCO test-devAPL70G-RMI
1 Image, 2*2 StitchiCOCO test-devAPM62.3G-RMI
1 Image, 2*2 StitchiCOCO test-devAR69.7G-RMI
1 Image, 2*2 StitchiCOCO test-devAR5088.7G-RMI
1 Image, 2*2 StitchiCOCO test-devAR7575.5G-RMI
1 Image, 2*2 StitchiCOCO test-devARL77.1G-RMI
1 Image, 2*2 StitchiCOCO test-devARM64.4G-RMI
1 Image, 2*2 StitchiCOCO test-challengeAP69.1G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeAP5085.9G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeAP7575.2G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeAPL82.4G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeAR75.1G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeAR5090.7G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeAR7580.7G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeARL74.5G-RMI*
1 Image, 2*2 StitchiCOCO test-challengeARM69.7G-RMI*
1 Image, 2*2 StitchiCOCO test-devAP64.9G-RMI
1 Image, 2*2 StitchiCOCO test-devAP5085.5G-RMI
1 Image, 2*2 StitchiCOCO test-devAP7571.3G-RMI
1 Image, 2*2 StitchiCOCO test-devAPL70G-RMI
1 Image, 2*2 StitchiCOCO test-devAPM62.3G-RMI
1 Image, 2*2 StitchiCOCO (Common Objects in Context)AP0.685G-RMI*
1 Image, 2*2 StitchiCOCO (Common Objects in Context)AP0.649G-RMI

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16