TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/TCPFormer: Learning Temporal Correlation with Implicit Pos...

TCPFormer: Learning Temporal Correlation with Implicit Pose Proxy for 3D Human Pose Estimation

Jiajie Liu, Mengyuan Liu, Hong Liu, Wenhao Li

2025-01-033D Human Pose EstimationMonocular 3D Human Pose EstimationPose Estimation
PaperPDFCode(official)

Abstract

Recent multi-frame lifting methods have dominated the 3D human pose estimation. However, previous methods ignore the intricate dependence within the 2D pose sequence and learn single temporal correlation. To alleviate this limitation, we propose TCPFormer, which leverages an implicit pose proxy as an intermediate representation. Each proxy within the implicit pose proxy can build one temporal correlation therefore helping us learn more comprehensive temporal correlation of human motion. Specifically, our method consists of three key components: Proxy Update Module (PUM), Proxy Invocation Module (PIM), and Proxy Attention Module (PAM). PUM first uses pose features to update the implicit pose proxy, enabling it to store representative information from the pose sequence. PIM then invocates and integrates the pose proxy with the pose sequence to enhance the motion semantics of each pose. Finally, PAM leverages the above mapping between the pose sequence and pose proxy to enhance the temporal correlation of the whole pose sequence. Experiments on the Human3.6M and MPI-INF-3DHP datasets demonstrate that our proposed TCPFormer outperforms the previous state-of-the-art methods.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationMPI-INF-3DHPAUC87.7TCPFormer (T=81)
3D Human Pose EstimationMPI-INF-3DHPMPJPE15TCPFormer (T=81)
3D Human Pose EstimationMPI-INF-3DHPPCK99TCPFormer (T=81)
3D Human Pose EstimationMPI-INF-3DHPAUC86.5TCPFormer (T=27)
3D Human Pose EstimationMPI-INF-3DHPMPJPE17.8TCPFormer (T=27)
3D Human Pose EstimationMPI-INF-3DHPPCK98.7TCPFormer (T=27)
Pose EstimationMPI-INF-3DHPAUC87.7TCPFormer (T=81)
Pose EstimationMPI-INF-3DHPMPJPE15TCPFormer (T=81)
Pose EstimationMPI-INF-3DHPPCK99TCPFormer (T=81)
Pose EstimationMPI-INF-3DHPAUC86.5TCPFormer (T=27)
Pose EstimationMPI-INF-3DHPMPJPE17.8TCPFormer (T=27)
Pose EstimationMPI-INF-3DHPPCK98.7TCPFormer (T=27)
3DMPI-INF-3DHPAUC87.7TCPFormer (T=81)
3DMPI-INF-3DHPMPJPE15TCPFormer (T=81)
3DMPI-INF-3DHPPCK99TCPFormer (T=81)
3DMPI-INF-3DHPAUC86.5TCPFormer (T=27)
3DMPI-INF-3DHPMPJPE17.8TCPFormer (T=27)
3DMPI-INF-3DHPPCK98.7TCPFormer (T=27)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC87.7TCPFormer (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE15TCPFormer (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK99TCPFormer (T=81)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC86.5TCPFormer (T=27)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE17.8TCPFormer (T=27)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK98.7TCPFormer (T=27)

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16