TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Explicit Occlusion Reasoning for Multi-person 3D Human Pos...

Explicit Occlusion Reasoning for Multi-person 3D Human Pose Estimation

Qihao Liu, Yi Zhang, Song Bai, Alan Yuille

2022-07-293D Human Pose EstimationData AugmentationPose Estimation3D Multi-Person Pose Estimation (root-relative)3D Multi-Person Pose Estimation (absolute)
PaperPDF

Abstract

Occlusion poses a great threat to monocular multi-person 3D human pose estimation due to large variability in terms of the shape, appearance, and position of occluders. While existing methods try to handle occlusion with pose priors/constraints, data augmentation, or implicit reasoning, they still fail to generalize to unseen poses or occlusion cases and may make large mistakes when multiple people are present. Inspired by the remarkable ability of humans to infer occluded joints from visible cues, we develop a method to explicitly model this process that significantly improves bottom-up multi-person human pose estimation with or without occlusions. First, we split the task into two subtasks: visible keypoints detection and occluded keypoints reasoning, and propose a Deeply Supervised Encoder Distillation (DSED) network to solve the second one. To train our model, we propose a Skeleton-guided human Shape Fitting (SSF) approach to generate pseudo occlusion labels on the existing datasets, enabling explicit occlusion reasoning. Experiments show that explicitly learning from occlusions improves human pose estimation. In addition, exploiting feature-level information of visible joints allows us to reason about occluded joints more accurately. Our method outperforms both the state-of-the-art top-down and bottom-up methods on several benchmarks.

Results

TaskDatasetMetricValueModel
3D Multi-Person Pose Estimation (root-relative)MuPoTS-3D3DPCK79.4Liu et al.
3D Human Pose EstimationMuPoTS-3D3DPCK36.5Liu et al.
3D Human Pose EstimationMuPoTS-3D3DPCK79.4Liu et al.
3D Multi-Person Pose Estimation (absolute)MuPoTS-3D3DPCK36.5Liu et al.
Pose EstimationMuPoTS-3D3DPCK36.5Liu et al.
Pose EstimationMuPoTS-3D3DPCK79.4Liu et al.
3DMuPoTS-3D3DPCK36.5Liu et al.
3DMuPoTS-3D3DPCK79.4Liu et al.
3D Multi-Person Pose EstimationMuPoTS-3D3DPCK36.5Liu et al.
3D Multi-Person Pose EstimationMuPoTS-3D3DPCK79.4Liu et al.
1 Image, 2*2 StitchiMuPoTS-3D3DPCK36.5Liu et al.
1 Image, 2*2 StitchiMuPoTS-3D3DPCK79.4Liu et al.

Related Papers

Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16