TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/RepNet: Weakly Supervised Training of an Adversarial Repro...

RepNet: Weakly Supervised Training of an Adversarial Reprojection Network for 3D Human Pose Estimation

Bastian Wandt, Bodo Rosenhahn

2019-02-26CVPR 2019 63D Human Pose EstimationWeakly-supervised 3D Human Pose EstimationMonocular 3D Human Pose EstimationPose EstimationMemorization
PaperPDFCode

Abstract

This paper addresses the problem of 3D human pose estimation from single images. While for a long time human skeletons were parameterized and fitted to the observation by satisfying a reprojection error, nowadays researchers directly use neural networks to infer the 3D pose from the observations. However, most of these approaches ignore the fact that a reprojection constraint has to be satisfied and are sensitive to overfitting. We tackle the overfitting problem by ignoring 2D to 3D correspondences. This efficiently avoids a simple memorization of the training data and allows for a weakly supervised training. One part of the proposed reprojection network (RepNet) learns a mapping from a distribution of 2D poses to a distribution of 3D poses using an adversarial training approach. Another part of the network estimates the camera. This allows for the definition of a network layer that performs the reprojection of the estimated 3D pose back to 2D which results in a reprojection loss function. Our experiments show that RepNet generalizes well to unknown data and outperforms state-of-the-art methods when applied to unseen data. Moreover, our implementation runs in real-time on a standard desktop PC.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationMPI-INF-3DHPAUC54.8RepNet (H36M)
3D Human Pose EstimationMPI-INF-3DHPMPJPE92.5RepNet (H36M)
3D Human Pose EstimationMPI-INF-3DHPPCK81.8RepNet (H36M)
3D Human Pose EstimationMPI-INF-3DHPAUC58.5RepNet (3DHP)
3D Human Pose EstimationMPI-INF-3DHPMPJPE97.8RepNet (3DHP)
3D Human Pose EstimationMPI-INF-3DHPPCK82.5RepNet (3DHP)
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)89.9RepNet
3D Human Pose EstimationHuman3.6MFrames Needed1RepNet
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)89.9RepNet
3D Human Pose EstimationHuman3.6MNumber of Frames Per View1RepNet
3D Human Pose EstimationHuman3.6MNumber of Views1RepNet
Pose EstimationMPI-INF-3DHPAUC54.8RepNet (H36M)
Pose EstimationMPI-INF-3DHPMPJPE92.5RepNet (H36M)
Pose EstimationMPI-INF-3DHPPCK81.8RepNet (H36M)
Pose EstimationMPI-INF-3DHPAUC58.5RepNet (3DHP)
Pose EstimationMPI-INF-3DHPMPJPE97.8RepNet (3DHP)
Pose EstimationMPI-INF-3DHPPCK82.5RepNet (3DHP)
Pose EstimationHuman3.6MAverage MPJPE (mm)89.9RepNet
Pose EstimationHuman3.6MFrames Needed1RepNet
Pose EstimationHuman3.6MAverage MPJPE (mm)89.9RepNet
Pose EstimationHuman3.6MNumber of Frames Per View1RepNet
Pose EstimationHuman3.6MNumber of Views1RepNet
3DMPI-INF-3DHPAUC54.8RepNet (H36M)
3DMPI-INF-3DHPMPJPE92.5RepNet (H36M)
3DMPI-INF-3DHPPCK81.8RepNet (H36M)
3DMPI-INF-3DHPAUC58.5RepNet (3DHP)
3DMPI-INF-3DHPMPJPE97.8RepNet (3DHP)
3DMPI-INF-3DHPPCK82.5RepNet (3DHP)
3DHuman3.6MAverage MPJPE (mm)89.9RepNet
3DHuman3.6MFrames Needed1RepNet
3DHuman3.6MAverage MPJPE (mm)89.9RepNet
3DHuman3.6MNumber of Frames Per View1RepNet
3DHuman3.6MNumber of Views1RepNet
1 Image, 2*2 StitchiMPI-INF-3DHPAUC54.8RepNet (H36M)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE92.5RepNet (H36M)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK81.8RepNet (H36M)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC58.5RepNet (3DHP)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE97.8RepNet (3DHP)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK82.5RepNet (3DHP)
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)89.9RepNet
1 Image, 2*2 StitchiHuman3.6MFrames Needed1RepNet
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)89.9RepNet
1 Image, 2*2 StitchiHuman3.6MNumber of Frames Per View1RepNet
1 Image, 2*2 StitchiHuman3.6MNumber of Views1RepNet

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16