TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Deep High-Resolution Representation Learning for Human Pos...

Deep High-Resolution Representation Learning for Human Pose Estimation

Ke Sun, Bin Xiao, Dong Liu, Jingdong Wang

2019-02-25CVPR 2019 63D Human Pose EstimationRepresentation Learning2D Human Pose EstimationVocal Bursts Intensity PredictionPose EstimationMulti-Person Pose EstimationKeypoint DetectionPose TrackingInstance Segmentation3D Pose Estimation2D Pose EstimationObject Detection
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCodeCode

Abstract

This is an official pytorch implementation of Deep High-Resolution Representation Learning for Human Pose Estimation. In this work, we are interested in the human pose estimation problem with a focus on learning reliable high-resolution representations. Most existing methods recover high-resolution representations from low-resolution representations produced by a high-to-low resolution network. Instead, our proposed network maintains high-resolution representations through the whole process. We start from a high-resolution subnetwork as the first stage, gradually add high-to-low resolution subnetworks one by one to form more stages, and connect the mutli-resolution subnetworks in parallel. We conduct repeated multi-scale fusions such that each of the high-to-low resolution representations receives information from other parallel representations over and over, leading to rich high-resolution representations. As a result, the predicted keypoint heatmap is potentially more accurate and spatially more precise. We empirically demonstrate the effectiveness of our network through the superior pose estimation results over two benchmark datasets: the COCO keypoint detection dataset and the MPII Human Pose dataset. The code and models have been publicly available at \url{https://github.com/leoxiaobin/deep-high-resolution-net.pytorch}.

Results

TaskDatasetMetricValueModel
Pose EstimationCOCO val2017AP75.3HRNet (256x192)
Pose EstimationAICAP33.5HRNet (HRNet-w48 )
Pose EstimationAICAP5078HRNet (HRNet-w48 )
Pose EstimationAICAP7523.6HRNet (HRNet-w48 )
Pose EstimationAICAR37.9HRNet (HRNet-w48 )
Pose EstimationAICAR5080HRNet (HRNet-w48 )
Pose EstimationAICAP32.3HRNet (HRNet-w32)
Pose EstimationAICAP5076.2HRNet (HRNet-w32)
Pose EstimationAICAP7521.9HRNet (HRNet-w32)
Pose EstimationAICAR36.6HRNet (HRNet-w32)
Pose EstimationAICAR5078.9HRNet (HRNet-w32)
Pose EstimationCOCO test-devAP77HRNet-W48 + extra data
Pose EstimationCOCO test-devAP5092.7HRNet-W48 + extra data
Pose EstimationCOCO test-devAP7584.5HRNet-W48 + extra data
Pose EstimationCOCO test-devAPL83.1HRNet-W48 + extra data
Pose EstimationCOCO test-devAPM73.4HRNet-W48 + extra data
Pose EstimationCOCO test-devAR82HRNet-W48 + extra data
Pose EstimationMPII Human PosePCKh-0.592.3HRNet-W32
Pose EstimationBRACEAverage Precision0.357HRNet fine-tuned on BRACE
Pose EstimationBRACEAverage Recall0.445HRNet fine-tuned on BRACE
Pose EstimationBRACEAverage Precision0.158HRNet pre-trained on COCO
Pose EstimationBRACEAverage Recall0.202HRNet pre-trained on COCO
Pose EstimationCOCO test-devAP5092.7HRNet*
Pose EstimationCOCO test-devAP7584.5HRNet*
Pose EstimationCOCO test-devAPL83.1HRNet*
Pose EstimationCOCO test-devAPM73.4HRNet*
Pose EstimationCOCO test-devAR82HRNet*
Pose EstimationCOCO test-devAP5092.5HRNet
Pose EstimationCOCO test-devAP7583.3HRNet
Pose EstimationCOCO test-devAPL81.5HRNet
Pose EstimationCOCO test-devAPM71.9HRNet
Pose EstimationCOCO test-devAR80.5HRNet
Pose EstimationCOCO (Common Objects in Context)Test AP75.5HRNet-48(384x288)
Pose EstimationCOCO (Common Objects in Context)Validation AP76.3HRNet-48(384x288)
Pose EstimationCOCO (Common Objects in Context)Validation AP75.8HRNet-32
Pose EstimationHARPERAverage MPJPE (mm)151HRNet + Depth
2D Pose EstimationHARPERPCK868HRNet
3DCOCO val2017AP75.3HRNet (256x192)
3DAICAP33.5HRNet (HRNet-w48 )
3DAICAP5078HRNet (HRNet-w48 )
3DAICAP7523.6HRNet (HRNet-w48 )
3DAICAR37.9HRNet (HRNet-w48 )
3DAICAR5080HRNet (HRNet-w48 )
3DAICAP32.3HRNet (HRNet-w32)
3DAICAP5076.2HRNet (HRNet-w32)
3DAICAP7521.9HRNet (HRNet-w32)
3DAICAR36.6HRNet (HRNet-w32)
3DAICAR5078.9HRNet (HRNet-w32)
3DCOCO test-devAP77HRNet-W48 + extra data
3DCOCO test-devAP5092.7HRNet-W48 + extra data
3DCOCO test-devAP7584.5HRNet-W48 + extra data
3DCOCO test-devAPL83.1HRNet-W48 + extra data
3DCOCO test-devAPM73.4HRNet-W48 + extra data
3DCOCO test-devAR82HRNet-W48 + extra data
3DMPII Human PosePCKh-0.592.3HRNet-W32
3DBRACEAverage Precision0.357HRNet fine-tuned on BRACE
3DBRACEAverage Recall0.445HRNet fine-tuned on BRACE
3DBRACEAverage Precision0.158HRNet pre-trained on COCO
3DBRACEAverage Recall0.202HRNet pre-trained on COCO
3DCOCO test-devAP5092.7HRNet*
3DCOCO test-devAP7584.5HRNet*
3DCOCO test-devAPL83.1HRNet*
3DCOCO test-devAPM73.4HRNet*
3DCOCO test-devAR82HRNet*
3DCOCO test-devAP5092.5HRNet
3DCOCO test-devAP7583.3HRNet
3DCOCO test-devAPL81.5HRNet
3DCOCO test-devAPM71.9HRNet
3DCOCO test-devAR80.5HRNet
3DCOCO (Common Objects in Context)Test AP75.5HRNet-48(384x288)
3DCOCO (Common Objects in Context)Validation AP76.3HRNet-48(384x288)
3DCOCO (Common Objects in Context)Validation AP75.8HRNet-32
3DHARPERAverage MPJPE (mm)151HRNet + Depth
Pose TrackingPoseTrack2017MOTA57.93HRNet-W48 COCO
Pose TrackingPoseTrack2017mAP74.95HRNet-W48 COCO
2D Human Pose EstimationHuman-ArtAP0.417HRNet-w48
2D Human Pose EstimationHuman-ArtAP (gt bbox)0.769HRNet-w48
2D Human Pose EstimationHuman-ArtAP0.399HRNet-w32
2D Human Pose EstimationHuman-ArtAP (gt bbox)0.754HRNet-w32
2D Human Pose EstimationCOCO-WholeBodyWB43.2HRNet
2D Human Pose EstimationCOCO-WholeBodybody65.9HRNet
2D Human Pose EstimationCOCO-WholeBodyface52.3HRNet
2D Human Pose EstimationCOCO-WholeBodyfoot31.4HRNet
2D Human Pose EstimationCOCO-WholeBodyhand30HRNet
3D Pose EstimationHARPERAverage MPJPE (mm)151HRNet + Depth
2D ClassificationHARPERPCK868HRNet
1 Image, 2*2 StitchiCOCO val2017AP75.3HRNet (256x192)
1 Image, 2*2 StitchiAICAP33.5HRNet (HRNet-w48 )
1 Image, 2*2 StitchiAICAP5078HRNet (HRNet-w48 )
1 Image, 2*2 StitchiAICAP7523.6HRNet (HRNet-w48 )
1 Image, 2*2 StitchiAICAR37.9HRNet (HRNet-w48 )
1 Image, 2*2 StitchiAICAR5080HRNet (HRNet-w48 )
1 Image, 2*2 StitchiAICAP32.3HRNet (HRNet-w32)
1 Image, 2*2 StitchiAICAP5076.2HRNet (HRNet-w32)
1 Image, 2*2 StitchiAICAP7521.9HRNet (HRNet-w32)
1 Image, 2*2 StitchiAICAR36.6HRNet (HRNet-w32)
1 Image, 2*2 StitchiAICAR5078.9HRNet (HRNet-w32)
1 Image, 2*2 StitchiCOCO test-devAP77HRNet-W48 + extra data
1 Image, 2*2 StitchiCOCO test-devAP5092.7HRNet-W48 + extra data
1 Image, 2*2 StitchiCOCO test-devAP7584.5HRNet-W48 + extra data
1 Image, 2*2 StitchiCOCO test-devAPL83.1HRNet-W48 + extra data
1 Image, 2*2 StitchiCOCO test-devAPM73.4HRNet-W48 + extra data
1 Image, 2*2 StitchiCOCO test-devAR82HRNet-W48 + extra data
1 Image, 2*2 StitchiMPII Human PosePCKh-0.592.3HRNet-W32
1 Image, 2*2 StitchiBRACEAverage Precision0.357HRNet fine-tuned on BRACE
1 Image, 2*2 StitchiBRACEAverage Recall0.445HRNet fine-tuned on BRACE
1 Image, 2*2 StitchiBRACEAverage Precision0.158HRNet pre-trained on COCO
1 Image, 2*2 StitchiBRACEAverage Recall0.202HRNet pre-trained on COCO
1 Image, 2*2 StitchiCOCO test-devAP5092.7HRNet*
1 Image, 2*2 StitchiCOCO test-devAP7584.5HRNet*
1 Image, 2*2 StitchiCOCO test-devAPL83.1HRNet*
1 Image, 2*2 StitchiCOCO test-devAPM73.4HRNet*
1 Image, 2*2 StitchiCOCO test-devAR82HRNet*
1 Image, 2*2 StitchiCOCO test-devAP5092.5HRNet
1 Image, 2*2 StitchiCOCO test-devAP7583.3HRNet
1 Image, 2*2 StitchiCOCO test-devAPL81.5HRNet
1 Image, 2*2 StitchiCOCO test-devAPM71.9HRNet
1 Image, 2*2 StitchiCOCO test-devAR80.5HRNet
1 Image, 2*2 StitchiCOCO (Common Objects in Context)Test AP75.5HRNet-48(384x288)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)Validation AP76.3HRNet-48(384x288)
1 Image, 2*2 StitchiCOCO (Common Objects in Context)Validation AP75.8HRNet-32
1 Image, 2*2 StitchiHARPERAverage MPJPE (mm)151HRNet + Depth

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17