TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MUG: Multi-human Graph Network for 3D Mesh Reconstruction ...

MUG: Multi-human Graph Network for 3D Mesh Reconstruction from 2D Pose

Chenyan Wu, Yandong Li, Xianfeng Tang, James Wang

2022-05-253D Human Pose Estimation3D Multi-Person Human Pose Estimation3D Multi-Person Pose Estimation
PaperPDF

Abstract

Reconstructing multi-human body mesh from a single monocular image is an important but challenging computer vision problem. In addition to the individual body mesh models, we need to estimate relative 3D positions among subjects to generate a coherent representation. In this work, through a single graph neural network, named MUG (Multi-hUman Graph network), we construct coherent multi-human meshes using only multi-human 2D pose as input. Compared with existing methods, which adopt a detection-style pipeline (i.e., extracting image features and then locating human instances and recovering body meshes from that) and suffer from the significant domain gap between lab-collected training datasets and in-the-wild testing datasets, our method benefits from the 2D pose which has a relatively consistent geometric property across datasets. Our method works like the following: First, to model the multi-human environment, it processes multi-human 2D poses and builds a novel heterogeneous graph, where nodes from different people and within one person are connected to capture inter-human interactions and draw the body geometry (i.e., skeleton and mesh structure). Second, it employs a dual-branch graph neural network structure -- one for predicting inter-human depth relation and the other one for predicting root-joint-relative mesh coordinates. Finally, the entire multi-human 3D meshes are constructed by combining the output from both branches. Extensive experiments demonstrate that MUG outperforms previous multi-human mesh estimation methods on standard 3D human benchmarks -- Panoptic, MuPoTS-3D and 3DPW.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationPanopticAverage MPJPE (mm)127.8MUG
3D Human Pose Estimation3DPWMPJPE87MUG
3D Human Pose Estimation3DPWMPVPE106.2MUG
3D Human Pose Estimation3DPWPA-MPJPE60.5MUG
3D Human Pose EstimationMuPoTS-3D3DPCK76.27MUG
Pose EstimationPanopticAverage MPJPE (mm)127.8MUG
Pose Estimation3DPWMPJPE87MUG
Pose Estimation3DPWMPVPE106.2MUG
Pose Estimation3DPWPA-MPJPE60.5MUG
Pose EstimationMuPoTS-3D3DPCK76.27MUG
3DPanopticAverage MPJPE (mm)127.8MUG
3D3DPWMPJPE87MUG
3D3DPWMPVPE106.2MUG
3D3DPWPA-MPJPE60.5MUG
3DMuPoTS-3D3DPCK76.27MUG
3D Multi-Person Pose EstimationMuPoTS-3D3DPCK76.27MUG
1 Image, 2*2 StitchiPanopticAverage MPJPE (mm)127.8MUG
1 Image, 2*2 Stitchi3DPWMPJPE87MUG
1 Image, 2*2 Stitchi3DPWMPVPE106.2MUG
1 Image, 2*2 Stitchi3DPWPA-MPJPE60.5MUG
1 Image, 2*2 StitchiMuPoTS-3D3DPCK76.27MUG

Related Papers

Systematic Comparison of Projection Methods for Monocular 3D Human Pose Estimation on Fisheye Images2025-06-24ExtPose: Robust and Coherent Pose Estimation by Extending ViTs2025-06-18PoseGRAF: Geometric-Reinforced Adaptive Fusion for Monocular 3D Human Pose Estimation2025-06-17Learning Pyramid-structured Long-range Dependencies for 3D Human Pose Estimation2025-06-03UPTor: Unified 3D Human Pose Dynamics and Trajectory Prediction for Human-Robot Interaction2025-05-20PoseBench3D: A Cross-Dataset Analysis Framework for 3D Human Pose Estimation2025-05-16HDiffTG: A Lightweight Hybrid Diffusion-Transformer-GCN Architecture for 3D Human Pose Estimation2025-05-07Continuous Normalizing Flows for Uncertainty-Aware Human Pose Estimation2025-05-04