TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Space-Time-Separable Graph Convolutional Network for Pose ...

Space-Time-Separable Graph Convolutional Network for Pose Forecasting

Theodoros Sofianos, Alessio Sampieri, Luca Franco, Fabio Galasso

2021-10-09ICCV 2021 10Human Pose ForecastingTime SeriesTime Series AnalysisSTS
PaperPDFCode(official)

Abstract

Human pose forecasting is a complex structured-data sequence-modelling task, which has received increasing attention, also due to numerous potential applications. Research has mainly addressed the temporal dimension as time series and the interaction of human body joints with a kinematic tree or by a graph. This has decoupled the two aspects and leveraged progress from the relevant fields, but it has also limited the understanding of the complex structural joint spatio-temporal dynamics of the human pose. Here we propose a novel Space-Time-Separable Graph Convolutional Network (STS-GCN) for pose forecasting. For the first time, STS-GCN models the human pose dynamics only with a graph convolutional network (GCN), including the temporal evolution and the spatial joint interaction within a single-graph framework, which allows the cross-talk of motion and spatial correlations. Concurrently, STS-GCN is the first space-time-separable GCN: the space-time graph connectivity is factored into space and time affinity matrices, which bottlenecks the space-time cross-talk, while enabling full joint-joint and time-time correlations. Both affinity matrices are learnt end-to-end, which results in connections substantially deviating from the standard kinematic tree and the linear-time time series. In experimental evaluation on three complex, recent and large-scale benchmarks, Human3.6M [Ionescu et al. TPAMI'14], AMASS [Mahmood et al. ICCV'19] and 3DPW [Von Marcard et al. ECCV'18], STS-GCN outperforms the state-of-the-art, surpassing the current best technique [Mao et al. ECCV'20] by over 32% in average at the most difficult long-term predictions, while only requiring 1.7% of its parameters. We explain the results qualitatively and illustrate the graph interactions by the factored joint-joint and time-time learnt graph connections. Our source code is available at: https://github.com/FraLuca/STSGCN

Results

TaskDatasetMetricValueModel
Pose EstimationHARPERAverage MPJPE (mm) @ 1000ms171STS-GCN
Pose EstimationHARPERAverage MPJPE (mm) @ 400ms120STS-GCN
Pose EstimationHARPERLast Frame MPJPE (mm) @ 1000ms260STS-GCN
Pose EstimationHARPERLast Frame MPJPE (mm) @ 400ms147STS-GCN
Pose EstimationAMASSAverage MPJPE (mm) 1000 msec45.5STS-GCN
Pose EstimationHuman3.6MAverage MPJPE (mm) @ 1000 ms117STS-GCN
Pose EstimationHuman3.6MAverage MPJPE (mm) @ 400ms65.8STS-GCN
Pose EstimationHuman3.6MMAR, walking, 1,000ms0.87STS-GCN
Pose EstimationHuman3.6MMAR, walking, 400ms0.55STS-GCN
Pose Estimation3DPWAverage MPJPE (mm) 1000 msec42.3STS-GCN
3DHARPERAverage MPJPE (mm) @ 1000ms171STS-GCN
3DHARPERAverage MPJPE (mm) @ 400ms120STS-GCN
3DHARPERLast Frame MPJPE (mm) @ 1000ms260STS-GCN
3DHARPERLast Frame MPJPE (mm) @ 400ms147STS-GCN
3DAMASSAverage MPJPE (mm) 1000 msec45.5STS-GCN
3DHuman3.6MAverage MPJPE (mm) @ 1000 ms117STS-GCN
3DHuman3.6MAverage MPJPE (mm) @ 400ms65.8STS-GCN
3DHuman3.6MMAR, walking, 1,000ms0.87STS-GCN
3DHuman3.6MMAR, walking, 400ms0.55STS-GCN
3D3DPWAverage MPJPE (mm) 1000 msec42.3STS-GCN
1 Image, 2*2 StitchiHARPERAverage MPJPE (mm) @ 1000ms171STS-GCN
1 Image, 2*2 StitchiHARPERAverage MPJPE (mm) @ 400ms120STS-GCN
1 Image, 2*2 StitchiHARPERLast Frame MPJPE (mm) @ 1000ms260STS-GCN
1 Image, 2*2 StitchiHARPERLast Frame MPJPE (mm) @ 400ms147STS-GCN
1 Image, 2*2 StitchiAMASSAverage MPJPE (mm) 1000 msec45.5STS-GCN
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm) @ 1000 ms117STS-GCN
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm) @ 400ms65.8STS-GCN
1 Image, 2*2 StitchiHuman3.6MMAR, walking, 1,000ms0.87STS-GCN
1 Image, 2*2 StitchiHuman3.6MMAR, walking, 400ms0.55STS-GCN
1 Image, 2*2 Stitchi3DPWAverage MPJPE (mm) 1000 msec42.3STS-GCN

Related Papers

MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17The Power of Architecture: Deep Dive into Transformer Architectures for Long-Term Time Series Forecasting2025-07-17Emergence of Functionally Differentiated Structures via Mutual Information Optimization in Recurrent Neural Networks2025-07-17Data Augmentation in Time Series Forecasting through Inverted Framework2025-07-15D3FL: Data Distribution and Detrending for Robust Federated Learning in Non-linear Time-series Data2025-07-15Towards Interpretable Time Series Foundation Models2025-07-10MoFE-Time: Mixture of Frequency Domain Experts for Time-Series Forecasting Models2025-07-09Foundation models for time series forecasting: Application in conformal prediction2025-07-09