TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SMPLer-X: Scaling Up Expressive Human Pose and Shape Estim...

SMPLer-X: Scaling Up Expressive Human Pose and Shape Estimation

Zhongang Cai, Wanqi Yin, Ailing Zeng, Chen Wei, Qingping Sun, Yanjun Wang, Hui En Pang, Haiyi Mei, Mingyuan Zhang, Lei Zhang, Chen Change Loy, Lei Yang, Ziwei Liu

2023-09-29NeurIPS 2023 113D Human Pose EstimationBenchmarking3D Human Reconstruction3D Multi-Person Mesh Recovery
PaperPDFCodeCode(official)

Abstract

Expressive human pose and shape estimation (EHPS) unifies body, hands, and face motion capture with numerous applications. Despite encouraging progress, current state-of-the-art methods still depend largely on a confined set of training datasets. In this work, we investigate scaling up EHPS towards the first generalist foundation model (dubbed SMPLer-X), with up to ViT-Huge as the backbone and training with up to 4.5M instances from diverse data sources. With big data and the large model, SMPLer-X exhibits strong performance across diverse test benchmarks and excellent transferability to even unseen environments. 1) For the data scaling, we perform a systematic investigation on 32 EHPS datasets, including a wide range of scenarios that a model trained on any single dataset cannot handle. More importantly, capitalizing on insights obtained from the extensive benchmarking process, we optimize our training scheme and select datasets that lead to a significant leap in EHPS capabilities. 2) For the model scaling, we take advantage of vision transformers to study the scaling law of model sizes in EHPS. Moreover, our finetuning strategy turn SMPLer-X into specialist models, allowing them to achieve further performance boosts. Notably, our foundation model SMPLer-X consistently delivers state-of-the-art results on seven benchmarks such as AGORA (107.2 mm NMVE), UBody (57.4 mm PVE), EgoBody (63.6 mm PVE), and EHF (62.3 mm PVE without finetuning). Homepage: https://caizhongang.github.io/projects/SMPLer-X/

Results

TaskDatasetMetricValueModel
ReconstructionEHFMPVPE62.4SMPLer-X
ReconstructionEHFPA V2V (mm), whole body37.1SMPLer-X
3D Human Pose Estimation3DPWMPJPE75.2SMPLer-X
3D Human Pose EstimationUBodyPA-PVE-All31.9SMPLer-X
3D Human Pose EstimationUBodyPA-PVE-Face2.8SMPLer-X
3D Human Pose EstimationUBodyPA-PVE-Hands10.3SMPLer-X
3D Human Pose EstimationUBodyPVE-All57.5SMPLer-X
3D Human Pose EstimationUBodyPVE-Face21.6SMPLer-X
3D Human Pose EstimationUBodyPVE-Hands40.2SMPLer-X
3D Human Pose EstimationAGORAB-NMVE68.3SMPLer-X
3D Human Pose EstimationAGORAF-MVE29.9SMPLer-X
3D Human Pose EstimationAGORAFB-MVE99.7SMPLer-X
3D Human Pose EstimationAGORAFB-NMVE107.2SMPLer-X
3D Human Pose EstimationAGORALH/RH-MVE39.3SMPLer-X
Pose Estimation3DPWMPJPE75.2SMPLer-X
Pose EstimationUBodyPA-PVE-All31.9SMPLer-X
Pose EstimationUBodyPA-PVE-Face2.8SMPLer-X
Pose EstimationUBodyPA-PVE-Hands10.3SMPLer-X
Pose EstimationUBodyPVE-All57.5SMPLer-X
Pose EstimationUBodyPVE-Face21.6SMPLer-X
Pose EstimationUBodyPVE-Hands40.2SMPLer-X
Pose EstimationAGORAB-NMVE68.3SMPLer-X
Pose EstimationAGORAF-MVE29.9SMPLer-X
Pose EstimationAGORAFB-MVE99.7SMPLer-X
Pose EstimationAGORAFB-NMVE107.2SMPLer-X
Pose EstimationAGORALH/RH-MVE39.3SMPLer-X
3D3DPWMPJPE75.2SMPLer-X
3DUBodyPA-PVE-All31.9SMPLer-X
3DUBodyPA-PVE-Face2.8SMPLer-X
3DUBodyPA-PVE-Hands10.3SMPLer-X
3DUBodyPVE-All57.5SMPLer-X
3DUBodyPVE-Face21.6SMPLer-X
3DUBodyPVE-Hands40.2SMPLer-X
3DAGORAB-NMVE68.3SMPLer-X
3DAGORAF-MVE29.9SMPLer-X
3DAGORAFB-MVE99.7SMPLer-X
3DAGORAFB-NMVE107.2SMPLer-X
3DAGORALH/RH-MVE39.3SMPLer-X
3D Multi-Person Pose EstimationAGORAB-NMVE68.3SMPLer-X
3D Multi-Person Pose EstimationAGORAF-MVE29.9SMPLer-X
3D Multi-Person Pose EstimationAGORAFB-MVE99.7SMPLer-X
3D Multi-Person Pose EstimationAGORAFB-NMVE107.2SMPLer-X
3D Multi-Person Pose EstimationAGORALH/RH-MVE39.3SMPLer-X
1 Image, 2*2 Stitchi3DPWMPJPE75.2SMPLer-X
1 Image, 2*2 StitchiUBodyPA-PVE-All31.9SMPLer-X
1 Image, 2*2 StitchiUBodyPA-PVE-Face2.8SMPLer-X
1 Image, 2*2 StitchiUBodyPA-PVE-Hands10.3SMPLer-X
1 Image, 2*2 StitchiUBodyPVE-All57.5SMPLer-X
1 Image, 2*2 StitchiUBodyPVE-Face21.6SMPLer-X
1 Image, 2*2 StitchiUBodyPVE-Hands40.2SMPLer-X
1 Image, 2*2 StitchiAGORAB-NMVE68.3SMPLer-X
1 Image, 2*2 StitchiAGORAF-MVE29.9SMPLer-X
1 Image, 2*2 StitchiAGORAFB-MVE99.7SMPLer-X
1 Image, 2*2 StitchiAGORAFB-NMVE107.2SMPLer-X
1 Image, 2*2 StitchiAGORALH/RH-MVE39.3SMPLer-X

Related Papers

Visual Place Recognition for Large-Scale UAV Applications2025-07-20Training Transformers with Enforced Lipschitz Constants2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16DCR: Quantifying Data Contamination in LLMs Evaluation2025-07-15A Multi-View High-Resolution Foot-Ankle Complex Point Cloud Dataset During Gait for Occlusion-Robust 3D Completion2025-07-15FLsim: A Modular and Library-Agnostic Simulation Framework for Federated Learning2025-07-15