TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Multi-Graph Convolution Network for Pose Forecasting

Multi-Graph Convolution Network for Pose Forecasting

Hongwei Ren, Yuhong Shi, Kewei Liang

2023-04-11Human Pose ForecastingPose Prediction
PaperPDF

Abstract

Recently, there has been a growing interest in predicting human motion, which involves forecasting future body poses based on observed pose sequences. This task is complex due to modeling spatial and temporal relationships. The most commonly used models for this task are autoregressive models, such as recurrent neural networks (RNNs) or variants, and Transformer Networks. However, RNNs have several drawbacks, such as vanishing or exploding gradients. Other researchers have attempted to solve the communication problem in the spatial dimension by integrating Graph Convolutional Networks (GCN) and Long Short-Term Memory (LSTM) models. These works deal with temporal and spatial information separately, which limits the effectiveness. To fix this problem, we propose a novel approach called the multi-graph convolution network (MGCN) for 3D human pose forecasting. This model simultaneously captures spatial and temporal information by introducing an augmented graph for pose sequences. Multiple frames give multiple parts, joined together in a single graph instance. Furthermore, we also explore the influence of natural structure and sequence-aware attention to our model. In our experimental evaluation of the large-scale benchmark datasets, Human3.6M, AMSS and 3DPW, MGCN outperforms the state-of-the-art in pose prediction.

Results

TaskDatasetMetricValueModel
Pose EstimationHuman3.6MAverage MPJPE (mm) @ 1000 ms72.9MGCN
Pose EstimationHuman3.6MAverage MPJPE (mm) @ 400ms34.4MGCN
3DHuman3.6MAverage MPJPE (mm) @ 1000 ms72.9MGCN
3DHuman3.6MAverage MPJPE (mm) @ 400ms34.4MGCN
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm) @ 1000 ms72.9MGCN
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm) @ 400ms34.4MGCN

Related Papers

EasyInsert: A Data-Efficient and Generalizable Insertion Policy2025-05-22Template-Guided 3D Molecular Pose Generation via Flow Matching and Differentiable Optimization2025-05-22UPTor: Unified 3D Human Pose Dynamics and Trajectory Prediction for Human-Robot Interaction2025-05-20Multi-Resolution Haar Network: Enhancing human motion prediction via Haar transform2025-05-19ZeroGrasp: Zero-Shot Shape Reconstruction Enabled Robotic Grasping2025-04-15FRAME: Floor-aligned Representation for Avatar Motion from Egocentric Video2025-03-29Flow-NeRF: Joint Learning of Geometry, Poses, and Dense Flow within Unified Neural Representations2025-03-13MarsLGPR: Mars Rover Localization with Ground Penetrating Radar2025-03-06