TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/On the Calibration of Human Pose Estimation

On the Calibration of Human Pose Estimation

Kerui Gu, Rongyu Chen, Angela Yao

2023-11-282D Human Pose EstimationPose EstimationHuman Mesh Recovery
PaperPDF

Abstract

Most 2D human pose estimation frameworks estimate keypoint confidence in an ad-hoc manner, using heuristics such as the maximum value of heatmaps. The confidence is part of the evaluation scheme, e.g., AP for the MSCOCO dataset, yet has been largely overlooked in the development of state-of-the-art methods. This paper takes the first steps in addressing miscalibration in pose estimation. From a calibration point of view, the confidence should be aligned with the pose accuracy. In practice, existing methods are poorly calibrated. We show, through theoretical analysis, why a miscalibration gap exists and how to narrow the gap. Simply predicting the instance size and adjusting the confidence function gives considerable AP improvements. Given the black-box nature of deep neural networks, however, it is not possible to fully close this gap with only closed-form adjustments. As such, we go one step further and learn network-specific adjustments by enforcing consistency between confidence and pose accuracy. Our proposed Calibrated ConfidenceNet (CCNet) is a light-weight post-hoc addition that improves AP by up to 1.4% on off-the-shelf pose estimation frameworks. Applied to the downstream task of mesh recovery, CCNet facilitates an additional 1.0mm decrease in 3D keypoint error.

Results

TaskDatasetMetricValueModel
Pose EstimationCOCO val2017AP78.1CCNet (ViTPose-B_GT-bbox_256x192)
Pose EstimationCOCO val2017AP5093.7CCNet (ViTPose-B_GT-bbox_256x192)
Pose EstimationCOCO val2017AP7585CCNet (ViTPose-B_GT-bbox_256x192)
Pose EstimationCOCO val2017AR80.4CCNet (ViTPose-B_GT-bbox_256x192)
3DCOCO val2017AP78.1CCNet (ViTPose-B_GT-bbox_256x192)
3DCOCO val2017AP5093.7CCNet (ViTPose-B_GT-bbox_256x192)
3DCOCO val2017AP7585CCNet (ViTPose-B_GT-bbox_256x192)
3DCOCO val2017AR80.4CCNet (ViTPose-B_GT-bbox_256x192)
1 Image, 2*2 StitchiCOCO val2017AP78.1CCNet (ViTPose-B_GT-bbox_256x192)
1 Image, 2*2 StitchiCOCO val2017AP5093.7CCNet (ViTPose-B_GT-bbox_256x192)
1 Image, 2*2 StitchiCOCO val2017AP7585CCNet (ViTPose-B_GT-bbox_256x192)
1 Image, 2*2 StitchiCOCO val2017AR80.4CCNet (ViTPose-B_GT-bbox_256x192)

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16