TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/HigherHRNet: Scale-Aware Representation Learning for Botto...

HigherHRNet: Scale-Aware Representation Learning for Bottom-Up Human Pose Estimation

Bowen Cheng, Bin Xiao, Jingdong Wang, Honghui Shi, Thomas S. Huang, Lei Zhang

2019-08-27CVPR 2020 6Representation Learning2D Human Pose EstimationPose EstimationMulti-Person Pose EstimationPose Prediction
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCode

Abstract

Bottom-up human pose estimation methods have difficulties in predicting the correct pose for small persons due to challenges in scale variation. In this paper, we present HigherHRNet: a novel bottom-up human pose estimation method for learning scale-aware representations using high-resolution feature pyramids. Equipped with multi-resolution supervision for training and multi-resolution aggregation for inference, the proposed approach is able to solve the scale variation challenge in bottom-up multi-person pose estimation and localize keypoints more precisely, especially for small person. The feature pyramid in HigherHRNet consists of feature map outputs from HRNet and upsampled higher-resolution outputs through a transposed convolution. HigherHRNet outperforms the previous best bottom-up method by 2.5% AP for medium person on COCO test-dev, showing its effectiveness in handling scale variation. Furthermore, HigherHRNet achieves new state-of-the-art result on COCO test-dev (70.5% AP) without using refinement or other post-processing techniques, surpassing all existing bottom-up methods. HigherHRNet even surpasses all top-down methods on CrowdPose test (67.6% AP), suggesting its robustness in crowded scene. The code and models are available at https://github.com/HRNet/Higher-HRNet-Human-Pose-Estimation.

Results

TaskDatasetMetricValueModel
Pose EstimationUAV-HumanmAP56.5HigherHRNet
Pose EstimationCOCO test-devAP70.5HigherHRNet (HR-Net-48)
Pose EstimationCOCO test-devAP5089.3HigherHRNet (HR-Net-48)
Pose EstimationCOCO test-devAP7577.2HigherHRNet (HR-Net-48)
Pose EstimationCOCO test-devAPL75.8HigherHRNet (HR-Net-48)
Pose EstimationCOCO test-devAPM66.6HigherHRNet (HR-Net-48)
Pose EstimationCrowdPoseAP Easy75.8HigherHRNet(HR-Net-48)
Pose EstimationCrowdPoseAP Hard58.9HigherHRNet(HR-Net-48)
Pose EstimationCrowdPoseAP Medium68.1HigherHRNet(HR-Net-48)
Pose EstimationCrowdPosemAP @0.5:0.9567.6HigherHRNet(HR-Net-48)
3DUAV-HumanmAP56.5HigherHRNet
3DCOCO test-devAP70.5HigherHRNet (HR-Net-48)
3DCOCO test-devAP5089.3HigherHRNet (HR-Net-48)
3DCOCO test-devAP7577.2HigherHRNet (HR-Net-48)
3DCOCO test-devAPL75.8HigherHRNet (HR-Net-48)
3DCOCO test-devAPM66.6HigherHRNet (HR-Net-48)
3DCrowdPoseAP Easy75.8HigherHRNet(HR-Net-48)
3DCrowdPoseAP Hard58.9HigherHRNet(HR-Net-48)
3DCrowdPoseAP Medium68.1HigherHRNet(HR-Net-48)
3DCrowdPosemAP @0.5:0.9567.6HigherHRNet(HR-Net-48)
Multi-Person Pose EstimationCOCO test-devAP70.5HigherHRNet (HR-Net-48)
Multi-Person Pose EstimationCOCO test-devAP5089.3HigherHRNet (HR-Net-48)
Multi-Person Pose EstimationCOCO test-devAP7577.2HigherHRNet (HR-Net-48)
Multi-Person Pose EstimationCOCO test-devAPL75.8HigherHRNet (HR-Net-48)
Multi-Person Pose EstimationCOCO test-devAPM66.6HigherHRNet (HR-Net-48)
Multi-Person Pose EstimationCrowdPoseAP Easy75.8HigherHRNet(HR-Net-48)
Multi-Person Pose EstimationCrowdPoseAP Hard58.9HigherHRNet(HR-Net-48)
Multi-Person Pose EstimationCrowdPoseAP Medium68.1HigherHRNet(HR-Net-48)
Multi-Person Pose EstimationCrowdPosemAP @0.5:0.9567.6HigherHRNet(HR-Net-48)
1 Image, 2*2 StitchiUAV-HumanmAP56.5HigherHRNet
1 Image, 2*2 StitchiCOCO test-devAP70.5HigherHRNet (HR-Net-48)
1 Image, 2*2 StitchiCOCO test-devAP5089.3HigherHRNet (HR-Net-48)
1 Image, 2*2 StitchiCOCO test-devAP7577.2HigherHRNet (HR-Net-48)
1 Image, 2*2 StitchiCOCO test-devAPL75.8HigherHRNet (HR-Net-48)
1 Image, 2*2 StitchiCOCO test-devAPM66.6HigherHRNet (HR-Net-48)
1 Image, 2*2 StitchiCrowdPoseAP Easy75.8HigherHRNet(HR-Net-48)
1 Image, 2*2 StitchiCrowdPoseAP Hard58.9HigherHRNet(HR-Net-48)
1 Image, 2*2 StitchiCrowdPoseAP Medium68.1HigherHRNet(HR-Net-48)
1 Image, 2*2 StitchiCrowdPosemAP @0.5:0.9567.6HigherHRNet(HR-Net-48)

Related Papers

Touch in the Wild: Learning Fine-Grained Manipulation with a Portable Visuo-Tactile Gripper2025-07-20Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Boosting Team Modeling through Tempo-Relational Representation Learning2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17