TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/LInKs "Lifting Independent Keypoints" -- Partial Pose Lift...

LInKs "Lifting Independent Keypoints" -- Partial Pose Lifting for Occlusion Handling with Improved Accuracy in 2D-3D Human Pose Estimation

Peter Hardy, Hansung Kim

2023-09-133D Human Pose EstimationDimensionality ReductionOcclusion HandlingAttributeUnsupervised 3D Human Pose EstimationPose Estimation
PaperPDF

Abstract

We present LInKs, a novel unsupervised learning method to recover 3D human poses from 2D kinematic skeletons obtained from a single image, even when occlusions are present. Our approach follows a unique two-step process, which involves first lifting the occluded 2D pose to the 3D domain, followed by filling in the occluded parts using the partially reconstructed 3D coordinates. This lift-then-fill approach leads to significantly more accurate results compared to models that complete the pose in 2D space alone. Additionally, we improve the stability and likelihood estimation of normalising flows through a custom sampling function replacing PCA dimensionality reduction previously used in prior work. Furthermore, we are the first to investigate if different parts of the 2D kinematic skeleton can be lifted independently which we find by itself reduces the error of current lifting approaches. We attribute this to the reduction of long-range keypoint correlations. In our detailed evaluation, we quantify the error under various realistic occlusion scenarios, showcasing the versatility and applicability of our model. Our results consistently demonstrate the superiority of handling all types of occlusions in 3D space when compared to others that complete the pose in 2D space. Our approach also exhibits consistent accuracy in scenarios without occlusion, as evidenced by a 7.9% reduction in reconstruction error compared to prior works on the Human3.6M dataset. Furthermore, our method excels in accurately retrieving complete 3D poses even in the presence of occlusions, making it highly applicable in situations where complete 2D pose information is unavailable.

Results

TaskDatasetMetricValueModel
3D ReconstructionMPI-INF-3DHPAUC54LInKs
3D ReconstructionMPI-INF-3DHPPA-MPJPE49.7LInKs
3D ReconstructionMPI-INF-3DHPPCK86.3LInKs
3D ReconstructionHuman3.6MN-MPJPE61.6LInKs
3D ReconstructionHuman3.6MPA-MPJPE33.8LInKs
3DMPI-INF-3DHPAUC54LInKs
3DMPI-INF-3DHPPA-MPJPE49.7LInKs
3DMPI-INF-3DHPPCK86.3LInKs
3DHuman3.6MN-MPJPE61.6LInKs
3DHuman3.6MPA-MPJPE33.8LInKs

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17MGFFD-VLM: Multi-Granularity Prompt Learning for Face Forgery Detection with VLM2025-07-16Non-Adaptive Adversarial Face Generation2025-07-16SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16