TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/PGformer: Proxy-Bridged Game Transformer for Multi-Person ...

PGformer: Proxy-Bridged Game Transformer for Multi-Person Highly Interactive Extreme Motion Prediction

Yanwen Fang, Jintai Chen, Peng-Tao Jiang, Chao Li, Yifeng Geng, Eddy K. F. LAM, Guodong Li

2023-06-06motion predictionMulti-Person Pose forecasting
PaperPDF

Abstract

Multi-person motion prediction is a challenging task, especially for real-world scenarios of highly interacted persons. Most previous works have been devoted to studying the case of weak interactions (e.g., walking together), in which typically forecasting each human pose in isolation can still achieve good performances. This paper focuses on collaborative motion prediction for multiple persons with extreme motions and attempts to explore the relationships between the highly interactive persons' pose trajectories. Specifically, a novel cross-query attention (XQA) module is proposed to bilaterally learn the cross-dependencies between the two pose sequences tailored for this situation. A proxy unit is additionally introduced to bridge the involved persons, which cooperates with our proposed XQA module and subtly controls the bidirectional spatial information flows. These designs are then integrated into a Transformer-based architecture and the resulting model is called Proxy-bridged Game Transformer (PGformer) for multi-person interactive motion prediction. Its effectiveness has been evaluated on the challenging ExPI dataset, which involves highly interactive actions. Our PGformer consistently outperforms the state-of-the-art methods in both short- and long-term predictions by a large margin. Besides, our approach can also be compatible with the weakly interacted CMU-Mocap and MuPoTS-3D datasets and extended to the case of more than 2 individuals with encouraging results.

Results

TaskDatasetMetricValueModel
Autonomous VehiclesExpi - common actions splitAverage MPJPE (mm) @ 1000 ms231PGformer
Autonomous VehiclesExpi - common actions splitAverage MPJPE (mm) @ 200 ms53PGformer
Autonomous VehiclesExpi - common actions splitAverage MPJPE (mm) @ 400 ms108PGformer
Autonomous VehiclesExpi - common actions splitAverage MPJPE (mm) @ 600 ms156PGformer
Motion ForecastingExpi - common actions splitAverage MPJPE (mm) @ 1000 ms231PGformer
Motion ForecastingExpi - common actions splitAverage MPJPE (mm) @ 200 ms53PGformer
Motion ForecastingExpi - common actions splitAverage MPJPE (mm) @ 400 ms108PGformer
Motion ForecastingExpi - common actions splitAverage MPJPE (mm) @ 600 ms156PGformer
Autonomous DrivingExpi - common actions splitAverage MPJPE (mm) @ 1000 ms231PGformer
Autonomous DrivingExpi - common actions splitAverage MPJPE (mm) @ 200 ms53PGformer
Autonomous DrivingExpi - common actions splitAverage MPJPE (mm) @ 400 ms108PGformer
Autonomous DrivingExpi - common actions splitAverage MPJPE (mm) @ 600 ms156PGformer

Related Papers

Stochastic Human Motion Prediction with Memory of Action Transition and Action Characteristic2025-07-05Temporal Continual Learning with Prior Compensation for Human Motion Prediction2025-07-05AMPLIFY: Actionless Motion Priors for Robot Learning from Videos2025-06-17FocalAD: Local Motion Planning for End-to-End Autonomous Driving2025-06-13TrajFlow: Multi-modal Motion Prediction via Flow Matching2025-06-10HUMOF: Human Motion Forecasting in Interactive Social Scenes2025-06-04Rodrigues Network for Learning Robot Actions2025-06-03Autoregression-free video prediction using diffusion model for mitigating error propagation2025-05-28