TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Trajectory Space Factorization for Deep Video-Based 3D Hum...

Trajectory Space Factorization for Deep Video-Based 3D Human Pose Estimation

Jiahao Lin, Gim Hee Lee

2019-08-223D Human Pose EstimationMonocular 3D Human Pose EstimationPose Estimation3D Pose Estimation
PaperPDFCode(official)

Abstract

Existing deep learning approaches on 3d human pose estimation for videos are either based on Recurrent or Convolutional Neural Networks (RNNs or CNNs). However, RNN-based frameworks can only tackle sequences with limited frames because sequential models are sensitive to bad frames and tend to drift over long sequences. Although existing CNN-based temporal frameworks attempt to address the sensitivity and drift problems by concurrently processing all input frames in the sequence, the existing state-of-the-art CNN-based framework is limited to 3d pose estimation of a single frame from a sequential input. In this paper, we propose a deep learning-based framework that utilizes matrix factorization for sequential 3d human poses estimation. Our approach processes all input frames concurrently to avoid the sensitivity and drift problems, and yet outputs the 3d pose estimates for every frame in the input sequence. More specifically, the 3d poses in all frames are represented as a motion matrix factorized into a trajectory bases matrix and a trajectory coefficient matrix. The trajectory bases matrix is precomputed from matrix factorization approaches such as Singular Value Decomposition (SVD) or Discrete Cosine Transform (DCT), and the problem of sequential 3d pose estimation is reduced to training a deep network to regress the trajectory coefficient matrix. We demonstrate the effectiveness of our framework on long sequences by achieving state-of-the-art performances on multiple benchmark datasets. Our source code is available at: https://github.com/jiahaoLjh/trajectory-pose-3d.

Results

TaskDatasetMetricValueModel
3D Human Pose EstimationMPI-INF-3DHPAUC51.4Trajectory Space Factorization (F=25)
3D Human Pose EstimationMPI-INF-3DHPMPJPE79.8Trajectory Space Factorization (F=25)
3D Human Pose EstimationMPI-INF-3DHPPCK83.6Trajectory Space Factorization (F=25)
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
3D Human Pose EstimationHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
3D Human Pose EstimationHuman3.6MFrames Needed50Trajectory Space Factorization (50 frames)
Pose EstimationLeeds Sports PosesPCK83.6Trajectory Space Factorization (F=25)
Pose EstimationMPI-INF-3DHPAUC51.4Trajectory Space Factorization (F=25)
Pose EstimationMPI-INF-3DHPMPJPE79.8Trajectory Space Factorization (F=25)
Pose EstimationMPI-INF-3DHPPCK83.6Trajectory Space Factorization (F=25)
Pose EstimationHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
Pose EstimationHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
Pose EstimationHuman3.6MFrames Needed50Trajectory Space Factorization (50 frames)
3DLeeds Sports PosesPCK83.6Trajectory Space Factorization (F=25)
3DMPI-INF-3DHPAUC51.4Trajectory Space Factorization (F=25)
3DMPI-INF-3DHPMPJPE79.8Trajectory Space Factorization (F=25)
3DMPI-INF-3DHPPCK83.6Trajectory Space Factorization (F=25)
3DHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
3DHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
3DHuman3.6MFrames Needed50Trajectory Space Factorization (50 frames)
1 Image, 2*2 StitchiLeeds Sports PosesPCK83.6Trajectory Space Factorization (F=25)
1 Image, 2*2 StitchiMPI-INF-3DHPAUC51.4Trajectory Space Factorization (F=25)
1 Image, 2*2 StitchiMPI-INF-3DHPMPJPE79.8Trajectory Space Factorization (F=25)
1 Image, 2*2 StitchiMPI-INF-3DHPPCK83.6Trajectory Space Factorization (F=25)
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
1 Image, 2*2 StitchiHuman3.6MAverage MPJPE (mm)46.6Trajectory Space Factorization (50 frames)
1 Image, 2*2 StitchiHuman3.6MFrames Needed50Trajectory Space Factorization (50 frames)

Related Papers

$π^3$: Scalable Permutation-Equivariant Visual Geometry Learning2025-07-17Revisiting Reliability in the Reasoning-based Pose Estimation Benchmark2025-07-17DINO-VO: A Feature-based Visual Odometry Leveraging a Visual Foundation Model2025-07-17From Neck to Head: Bio-Impedance Sensing for Head Pose Estimation2025-07-17AthleticsPose: Authentic Sports Motion Dataset on Athletic Field and Evaluation of Monocular 3D Pose Estimation Ability2025-07-17SpatialTrackerV2: 3D Point Tracking Made Easy2025-07-16SGLoc: Semantic Localization System for Camera Pose Estimation from 3D Gaussian Splatting Representation2025-07-16Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16