TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Monocular Expressive Body Regression through Body-Driven A...

Monocular Expressive Body Regression through Body-Driven Attention

Vasileios Choutas, Georgios Pavlakos, Timo Bolkart, Dimitrios Tzionas, Michael J. Black

2020-08-20ECCV 2020 83D Human Pose Estimation3D Hand Pose Estimationregression3D Human Reconstruction3D Face Reconstruction3D Multi-Person Mesh Recovery
PaperPDFCode(official)

Abstract

To understand how people look, interact, or perform tasks, we need to quickly and accurately capture their 3D body, face, and hands together from an RGB image. Most existing methods focus only on parts of the body. A few recent approaches reconstruct full expressive 3D humans from images using 3D body models that include the face and hands. These methods are optimization-based and thus slow, prone to local optima, and require 2D keypoints as input. We address these limitations by introducing ExPose (EXpressive POse and Shape rEgression), which directly regresses the body, face, and hands, in SMPL-X format, from an RGB image. This is a hard problem due to the high dimensionality of the body and the lack of expressive training data. Additionally, hands and faces are much smaller than the body, occupying very few image pixels. This makes hand and face estimation hard when body images are downscaled for neural networks. We make three main contributions. First, we account for the lack of training data by curating a dataset of SMPL-X fits on in-the-wild images. Second, we observe that body estimation localizes the face and hands reasonably well. We introduce body-driven attention for face and hand regions in the original image to extract higher-resolution crops that are fed to dedicated refinement modules. Third, these modules exploit part-specific knowledge from existing face- and hand-only datasets. ExPose estimates expressive 3D humans more accurately than existing optimization methods at a small fraction of the computational cost. Our data, model and code are available for research at https://expose.is.tue.mpg.de .

Results

TaskDatasetMetricValueModel
ReconstructionExpressive hands and faces dataset (EHF)MPJPE, left hand13.5ExPose
ReconstructionExpressive hands and faces dataset (EHF)MPJPE-1462.8ExPose
ReconstructionExpressive hands and faces dataset (EHF)PA V2V (mm), body only52.6ExPose
ReconstructionExpressive hands and faces dataset (EHF)PA V2V (mm), face5.8ExPose
ReconstructionExpressive hands and faces dataset (EHF)PA V2V (mm), left hand13.1ExPose
ReconstructionExpressive hands and faces dataset (EHF)PA V2V (mm), whole body54.5ExPose
ReconstructionExpressive hands and faces dataset (EHF)TR V2V (mm), body only76.8ExPose
ReconstructionExpressive hands and faces dataset (EHF)TR V2V (mm), face15.9ExPose
ReconstructionExpressive hands and faces dataset (EHF)TR V2V (mm), left hand31.2ExPose
ReconstructionExpressive hands and faces dataset (EHF)TR V2V (mm), whole body65.7ExPose
ReconstructionExpressive hands and faces dataset (EHF)mean P2S28.9ExPose
ReconstructionExpressive hands and faces dataset (EHF)median P2S18ExPose
ReconstructionAGORAB-MPJPE150.4ExPose
ReconstructionAGORAB-MVE151.5ExPose
ReconstructionAGORAB-NMJE183.4ExPose
ReconstructionAGORAB-NMVE184.8ExPose
ReconstructionAGORAF-MPJPE55.2ExPose
ReconstructionAGORAF-MVE51.1ExPose
ReconstructionAGORAFB-MPJPE215.9ExPose
ReconstructionAGORAFB-MVE217.3ExPose
ReconstructionAGORAFB-NMJE263.3ExPose
ReconstructionAGORAFB-NMVE265ExPose
ReconstructionExpressive hands and faces dataset (EHF).All54.5PA-V2V (mm)
3D Human Pose Estimation3DPWMPJPE93.4ExPose
3D Human Pose Estimation3DPWPA-MPJPE60.7ExPose
3D Human Pose EstimationAGORAB-MPJPE150.4ExPose
3D Human Pose EstimationAGORAB-MVE151.5ExPose
3D Human Pose EstimationAGORAB-NMJE183.4ExPose
3D Human Pose EstimationAGORAB-NMVE184.8ExPose
3D Human Pose EstimationAGORAF-MPJPE55.2ExPose
3D Human Pose EstimationAGORAF-MVE51.1ExPose
3D Human Pose EstimationAGORAFB-MPJPE215.9ExPose
3D Human Pose EstimationAGORAFB-MVE217.3ExPose
3D Human Pose EstimationAGORAFB-NMJE263.3ExPose
3D Human Pose EstimationAGORAFB-NMVE265ExPose
HandFreiHANDPA-F@15mm0.918ExPose (hand sub-network h)
HandFreiHANDPA-F@5mm0.484ExPose (hand sub-network h)
HandFreiHANDPA-MPJPE12.2ExPose (hand sub-network h)
HandFreiHANDPA-MPVPE11.8ExPose (hand sub-network h)
Pose Estimation3DPWMPJPE93.4ExPose
Pose Estimation3DPWPA-MPJPE60.7ExPose
Pose EstimationAGORAB-MPJPE150.4ExPose
Pose EstimationAGORAB-MVE151.5ExPose
Pose EstimationAGORAB-NMJE183.4ExPose
Pose EstimationAGORAB-NMVE184.8ExPose
Pose EstimationAGORAF-MPJPE55.2ExPose
Pose EstimationAGORAF-MVE51.1ExPose
Pose EstimationAGORAFB-MPJPE215.9ExPose
Pose EstimationAGORAFB-MVE217.3ExPose
Pose EstimationAGORAFB-NMJE263.3ExPose
Pose EstimationAGORAFB-NMVE265ExPose
Pose EstimationFreiHANDPA-F@15mm0.918ExPose (hand sub-network h)
Pose EstimationFreiHANDPA-F@5mm0.484ExPose (hand sub-network h)
Pose EstimationFreiHANDPA-MPJPE12.2ExPose (hand sub-network h)
Pose EstimationFreiHANDPA-MPVPE11.8ExPose (hand sub-network h)
Hand Pose EstimationFreiHANDPA-F@15mm0.918ExPose (hand sub-network h)
Hand Pose EstimationFreiHANDPA-F@5mm0.484ExPose (hand sub-network h)
Hand Pose EstimationFreiHANDPA-MPJPE12.2ExPose (hand sub-network h)
Hand Pose EstimationFreiHANDPA-MPVPE11.8ExPose (hand sub-network h)
3D3DPWMPJPE93.4ExPose
3D3DPWPA-MPJPE60.7ExPose
3DAGORAB-MPJPE150.4ExPose
3DAGORAB-MVE151.5ExPose
3DAGORAB-NMJE183.4ExPose
3DAGORAB-NMVE184.8ExPose
3DAGORAF-MPJPE55.2ExPose
3DAGORAF-MVE51.1ExPose
3DAGORAFB-MPJPE215.9ExPose
3DAGORAFB-MVE217.3ExPose
3DAGORAFB-NMJE263.3ExPose
3DAGORAFB-NMVE265ExPose
3DFreiHANDPA-F@15mm0.918ExPose (hand sub-network h)
3DFreiHANDPA-F@5mm0.484ExPose (hand sub-network h)
3DFreiHANDPA-MPJPE12.2ExPose (hand sub-network h)
3DFreiHANDPA-MPVPE11.8ExPose (hand sub-network h)
3D Multi-Person Pose EstimationAGORAB-MPJPE150.4ExPose
3D Multi-Person Pose EstimationAGORAB-MVE151.5ExPose
3D Multi-Person Pose EstimationAGORAB-NMJE183.4ExPose
3D Multi-Person Pose EstimationAGORAB-NMVE184.8ExPose
3D Multi-Person Pose EstimationAGORAF-MPJPE55.2ExPose
3D Multi-Person Pose EstimationAGORAF-MVE51.1ExPose
3D Multi-Person Pose EstimationAGORAFB-MPJPE215.9ExPose
3D Multi-Person Pose EstimationAGORAFB-MVE217.3ExPose
3D Multi-Person Pose EstimationAGORAFB-NMJE263.3ExPose
3D Multi-Person Pose EstimationAGORAFB-NMVE265ExPose
3D Hand Pose EstimationFreiHANDPA-F@15mm0.918ExPose (hand sub-network h)
3D Hand Pose EstimationFreiHANDPA-F@5mm0.484ExPose (hand sub-network h)
3D Hand Pose EstimationFreiHANDPA-MPJPE12.2ExPose (hand sub-network h)
3D Hand Pose EstimationFreiHANDPA-MPVPE11.8ExPose (hand sub-network h)
1 Image, 2*2 Stitchi3DPWMPJPE93.4ExPose
1 Image, 2*2 Stitchi3DPWPA-MPJPE60.7ExPose
1 Image, 2*2 StitchiAGORAB-MPJPE150.4ExPose
1 Image, 2*2 StitchiAGORAB-MVE151.5ExPose
1 Image, 2*2 StitchiAGORAB-NMJE183.4ExPose
1 Image, 2*2 StitchiAGORAB-NMVE184.8ExPose
1 Image, 2*2 StitchiAGORAF-MPJPE55.2ExPose
1 Image, 2*2 StitchiAGORAF-MVE51.1ExPose
1 Image, 2*2 StitchiAGORAFB-MPJPE215.9ExPose
1 Image, 2*2 StitchiAGORAFB-MVE217.3ExPose
1 Image, 2*2 StitchiAGORAFB-NMJE263.3ExPose
1 Image, 2*2 StitchiAGORAFB-NMVE265ExPose
1 Image, 2*2 StitchiFreiHANDPA-F@15mm0.918ExPose (hand sub-network h)
1 Image, 2*2 StitchiFreiHANDPA-F@5mm0.484ExPose (hand sub-network h)
1 Image, 2*2 StitchiFreiHANDPA-MPJPE12.2ExPose (hand sub-network h)
1 Image, 2*2 StitchiFreiHANDPA-MPVPE11.8ExPose (hand sub-network h)

Related Papers

Language Integration in Fine-Tuning Multimodal Large Language Models for Image-Based Regression2025-07-20Neural Network-Guided Symbolic Regression for Interpretable Descriptor Discovery in Perovskite Catalysts2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16Second-Order Bounds for [0,1]-Valued Regression via Betting Loss2025-07-16Sparse Regression Codes exploit Multi-User Diversity without CSI2025-07-15Bradley-Terry and Multi-Objective Reward Modeling Are Complementary2025-07-10Active Learning for Manifold Gaussian Process Regression2025-06-26A Survey of Predictive Maintenance Methods: An Analysis of Prognostics via Classification and Regression2025-06-25