TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Humans in 4D: Reconstructing and Tracking Humans with Tran...

Humans in 4D: Reconstructing and Tracking Humans with Transformers

Shubham Goel, Georgios Pavlakos, Jathushan Rajasegaran, Angjoo Kanazawa, Jitendra Malik

2023-05-31ICCV 2023 13D Human Pose EstimationHuman Mesh RecoveryPose TrackingAction Recognition
PaperPDFCode(official)

Abstract

We present an approach to reconstruct humans and track them over time. At the core of our approach, we propose a fully "transformerized" version of a network for human mesh recovery. This network, HMR 2.0, advances the state of the art and shows the capability to analyze unusual poses that have in the past been difficult to reconstruct from single images. To analyze video, we use 3D reconstructions from HMR 2.0 as input to a tracking system that operates in 3D. This enables us to deal with multiple people and maintain identities through occlusion events. Our complete approach, 4DHumans, achieves state-of-the-art results for tracking people from monocular video. Furthermore, we demonstrate the effectiveness of HMR 2.0 on the downstream task of action recognition, achieving significant improvements over previous pose-based action recognition approaches. Our code and models are available on the project website: https://shubham-goel.github.io/4dhumans/.

Results

TaskDatasetMetricValueModel
3D Human Pose Estimation3DPWMPJPE69.8HMR 2.0
3D Human Pose Estimation3DPWMPVPE82.2HMR 2.0
3D Human Pose Estimation3DPWPA-MPJPE44.4HMR 2.0
Pose Estimation3DPWMPJPE69.8HMR 2.0
Pose Estimation3DPWMPVPE82.2HMR 2.0
Pose Estimation3DPWPA-MPJPE44.4HMR 2.0
3D3DPWMPJPE69.8HMR 2.0
3D3DPWMPVPE82.2HMR 2.0
3D3DPWPA-MPJPE44.4HMR 2.0
Pose TrackingPoseTrack2018IDF179.34DHumans + ViTDet
Pose TrackingPoseTrack2018IDs3674DHumans + ViTDet
Pose TrackingPoseTrack2018MOTA61.94DHumans + ViTDet
1 Image, 2*2 Stitchi3DPWMPJPE69.8HMR 2.0
1 Image, 2*2 Stitchi3DPWMPVPE82.2HMR 2.0
1 Image, 2*2 Stitchi3DPWPA-MPJPE44.4HMR 2.0

Related Papers

From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17Failure Forecasting Boosts Robustness of Sim2Real Rhythmic Insertion Policies2025-07-09Robotic Manipulation by Imitating Generated Videos Without Physical Demonstrations2025-07-01Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25