TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Multiscale Residual Learning of Graph Convolutional Sequen...

Multiscale Residual Learning of Graph Convolutional Sequence Chunks for Human Motion Prediction

Mohsen Zand, Ali Etemad, Michael Greenspan

2023-08-31Human Pose ForecastingHuman motion predictionmotion predictionPrediction
PaperPDFCode

Abstract

A new method is proposed for human motion prediction by learning temporal and spatial dependencies. Recently, multiscale graphs have been developed to model the human body at higher abstraction levels, resulting in more stable motion prediction. Current methods however predetermine scale levels and combine spatially proximal joints to generate coarser scales based on human priors, even though movement patterns in different motion sequences vary and do not fully comply with a fixed graph of spatially connected joints. Another problem with graph convolutional methods is mode collapse, in which predicted poses converge around a mean pose with no discernible movements, particularly in long-term predictions. To tackle these issues, we propose ResChunk, an end-to-end network which explores dynamically correlated body components based on the pairwise relationships between all joints in individual sequences. ResChunk is trained to learn the residuals between target sequence chunks in an autoregressive manner to enforce the temporal connectivities between consecutive chunks. It is hence a sequence-to-sequence prediction network which considers dynamic spatio-temporal features of sequences at multiple levels. Our experiments on two challenging benchmark datasets, CMU Mocap and Human3.6M, demonstrate that our proposed method is able to effectively model the sequence information for motion prediction and outperform other techniques to set a new state-of-the-art. Our code is available at https://github.com/MohsenZand/ResChunk.

Results

TaskDatasetMetricValueModel
Pose EstimationHuman3.6MAverage MPJPE (mm) @ 1000 ms81.95ResChunk
Pose EstimationHuman3.6MAverage MPJPE (mm) @ 400ms56.41ResChunk
3DHuman3.6MAverage MPJPE (mm) @ 1000 ms81.95ResChunk
3DHuman3.6MAverage MPJPE (mm) @ 400ms56.41ResChunk
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm) @ 1000 ms81.95ResChunk
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm) @ 400ms56.41ResChunk

Related Papers

Multi-Strategy Improved Snake Optimizer Accelerated CNN-LSTM-Attention-Adaboost for Trajectory Prediction2025-07-21Generative Click-through Rate Prediction with Applications to Search Advertising2025-07-15Conformation-Aware Structure Prediction of Antigen-Recognizing Immune Proteins2025-07-11Foundation models for time series forecasting: Application in conformal prediction2025-07-09Predicting Graph Structure via Adapted Flux Balance Analysis2025-07-08Speech Quality Assessment Model Based on Mixture of Experts: System-Level Performance Enhancement and Utterance-Level Challenge Analysis2025-07-08A Wireless Foundation Model for Multi-Task Prediction2025-07-08High Order Collaboration-Oriented Federated Graph Neural Network for Accurate QoS Prediction2025-07-07