TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Occluded Human Body Capture with Self-Supervised Spatial-T...

Occluded Human Body Capture with Self-Supervised Spatial-Temporal Motion Prior

Buzhen Huang, Yuan Shu, Jingyi Ju, Yangang Wang

2022-07-123D Human Pose EstimationUnsupervised 3D Human Pose Estimation
PaperPDFCode(official)

Abstract

Although significant progress has been achieved on monocular maker-less human motion capture in recent years, it is still hard for state-of-the-art methods to obtain satisfactory results in occlusion scenarios. There are two main reasons: the one is that the occluded motion capture is inherently ambiguous as various 3D poses can map to the same 2D observations, which always results in an unreliable estimation. The other is that no sufficient occluded human data can be used for training a robust model. To address the obstacles, our key-idea is to employ non-occluded human data to learn a joint-level spatial-temporal motion prior for occluded human with a self-supervised strategy. To further reduce the gap between synthetic and real occlusion data, we build the first 3D occluded motion dataset~(OcMotion), which can be used for both training and testing. We encode the motions in 2D maps and synthesize occlusions on non-occluded data for the self-supervised training. A spatial-temporal layer is then designed to learn joint-level correlations. The learned prior reduces the ambiguities of occlusions and is robust to diverse occlusion types, which is then adopted to assist the occluded human motion capture. Experimental results show that our method can generate accurate and coherent human motions from occluded videos with good generalization ability and runtime efficiency. The dataset and code are publicly available at \url{https://github.com/boycehbz/CHOMP}.

Results

TaskDatasetMetricValueModel
3D Human Pose Estimation3DPWMPJPE83.7CHOMP
3D Human Pose Estimation3DPWMPVPE110.1CHOMP
3D Human Pose Estimation3DPWPA-MPJPE51.7CHOMP
3D ReconstructionHuman3.6MMPJPE64.7CHOMP
3D ReconstructionHuman3.6MPA-MPJPE40.1CHOMP
Pose Estimation3DPWMPJPE83.7CHOMP
Pose Estimation3DPWMPVPE110.1CHOMP
Pose Estimation3DPWPA-MPJPE51.7CHOMP
3D3DPWMPJPE83.7CHOMP
3D3DPWMPVPE110.1CHOMP
3D3DPWPA-MPJPE51.7CHOMP
3DHuman3.6MMPJPE64.7CHOMP
3DHuman3.6MPA-MPJPE40.1CHOMP
1 Image, 2*2 Stitchi3DPWMPJPE83.7CHOMP
1 Image, 2*2 Stitchi3DPWMPVPE110.1CHOMP
1 Image, 2*2 Stitchi3DPWPA-MPJPE51.7CHOMP

Related Papers

Systematic Comparison of Projection Methods for Monocular 3D Human Pose Estimation on Fisheye Images2025-06-24ExtPose: Robust and Coherent Pose Estimation by Extending ViTs2025-06-18PoseGRAF: Geometric-Reinforced Adaptive Fusion for Monocular 3D Human Pose Estimation2025-06-17Learning Pyramid-structured Long-range Dependencies for 3D Human Pose Estimation2025-06-03UPTor: Unified 3D Human Pose Dynamics and Trajectory Prediction for Human-Robot Interaction2025-05-20PoseBench3D: A Cross-Dataset Analysis Framework for 3D Human Pose Estimation2025-05-16HDiffTG: A Lightweight Hybrid Diffusion-Transformer-GCN Architecture for 3D Human Pose Estimation2025-05-07Continuous Normalizing Flows for Uncertainty-Aware Human Pose Estimation2025-05-04