TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Estimating Egocentric 3D Human Pose in the Wild with Exter...

Estimating Egocentric 3D Human Pose in the Wild with External Weak Supervision

Jian Wang, Lingjie Liu, Weipeng Xu, Kripasindhu Sarkar, Diogo Luvizon, Christian Theobalt

2022-01-20CVPR 2022 13D Human Pose EstimationEgocentric Pose EstimationPose Estimation
PaperPDF

Abstract

Egocentric 3D human pose estimation with a single fisheye camera has drawn a significant amount of attention recently. However, existing methods struggle with pose estimation from in-the-wild images, because they can only be trained on synthetic data due to the unavailability of large-scale in-the-wild egocentric datasets. Furthermore, these methods easily fail when the body parts are occluded by or interacting with the surrounding scene. To address the shortage of in-the-wild data, we collect a large-scale in-the-wild egocentric dataset called Egocentric Poses in the Wild (EgoPW). This dataset is captured by a head-mounted fisheye camera and an auxiliary external camera, which provides an additional observation of the human body from a third-person perspective during training. We present a new egocentric pose estimation method, which can be trained on the new dataset with weak external supervision. Specifically, we first generate pseudo labels for the EgoPW dataset with a spatio-temporal optimization method by incorporating the external-view supervision. The pseudo labels are then used to train an egocentric pose estimation network. To facilitate the network training, we propose a novel learning strategy to supervise the egocentric features with the high-quality features extracted by a pretrained external-view pose estimation model. The experiments show that our method predicts accurate 3D poses from a single in-the-wild egocentric image and outperforms the state-of-the-art methods both quantitatively and qualitatively.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationGlobalEgoMocap Test DatasetAverage MPJPE (mm)81.71EgoPW
3D Human Pose EstimationGlobalEgoMocap Test DatasetPA-MPJPE64.87EgoPW
3D Human Pose EstimationSceneEgoAverage MPJPE (mm)189.6EgoPW
3D Human Pose EstimationSceneEgoPA-MPJPE105.3EgoPW
Pose EstimationGlobalEgoMocap Test DatasetAverage MPJPE (mm)81.71EgoPW
Pose EstimationGlobalEgoMocap Test DatasetPA-MPJPE64.87EgoPW
Pose EstimationSceneEgoAverage MPJPE (mm)189.6EgoPW
Pose EstimationSceneEgoPA-MPJPE105.3EgoPW
3DGlobalEgoMocap Test DatasetAverage MPJPE (mm)81.71EgoPW
3DGlobalEgoMocap Test DatasetPA-MPJPE64.87EgoPW
3DSceneEgoAverage MPJPE (mm)189.6EgoPW
3DSceneEgoPA-MPJPE105.3EgoPW
1 Image, 2*2 StitchiGlobalEgoMocap Test DatasetAverage MPJPE (mm)81.71EgoPW
1 Image, 2*2 StitchiGlobalEgoMocap Test DatasetPA-MPJPE64.87EgoPW
1 Image, 2*2 StitchiSceneEgoAverage MPJPE (mm)189.6EgoPW
1 Image, 2*2 StitchiSceneEgoPA-MPJPE105.3EgoPW

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16