TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Sparseness Meets Deepness: 3D Human Pose Estimation from M...

Sparseness Meets Deepness: 3D Human Pose Estimation from Monocular Video

Xiaowei Zhou, Menglong Zhu, Spyridon Leonardos, Kosta Derpanis, Kostas Daniilidis

2015-11-30CVPR 2016 63D Human Pose EstimationMonocular 3D Human Pose EstimationPose Estimation3D Pose Estimation2D Pose Estimation
PaperPDFCode

Abstract

This paper addresses the challenge of 3D full-body human pose estimation from a monocular image sequence. Here, two cases are considered: (i) the image locations of the human joints are provided and (ii) the image locations of joints are unknown. In the former case, a novel approach is introduced that integrates a sparsity-driven 3D geometric prior and temporal smoothness. In the latter case, the former case is extended by treating the image locations of the joints as latent variables. A deep fully convolutional network is trained to predict the uncertainty maps of the 2D joint locations. The 3D pose estimates are realized via an Expectation-Maximization algorithm over the entire sequence, where it is shown that the 2D joint location uncertainties can be conveniently marginalized out during inference. Empirical evaluation on the Human3.6M dataset shows that the proposed approaches achieve greater 3D pose estimation accuracy over state-of-the-art baselines. Further, the proposed approach outperforms a publicly available 2D pose estimation baseline on the challenging PennAction dataset.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)113.01Sparseness Meets Deepness
3D Human Pose EstimationHuman3.6MFrames Needed300Sparseness Meets Deepness
Pose EstimationHuman3.6MAverage MPJPE (mm)113.01Sparseness Meets Deepness
Pose EstimationHuman3.6MFrames Needed300Sparseness Meets Deepness
3DHuman3.6MAverage MPJPE (mm)113.01Sparseness Meets Deepness
3DHuman3.6MFrames Needed300Sparseness Meets Deepness
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)113.01Sparseness Meets Deepness
1 Image, 2*2 StitchiHuman3.6MFrames Needed300Sparseness Meets Deepness

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16