TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Part-based Graph Convolutional Network for Action Recognit...

Part-based Graph Convolutional Network for Action Recognition

Kalpit Thakkar, P. J. Narayanan

2018-09-13Skeleton Based Action RecognitionAction RecognitionTemporal Action Localization
PaperPDFCode(official)

Abstract

Human actions comprise of joint motion of articulated body parts or `gestures'. Human skeleton is intuitively represented as a sparse graph with joints as nodes and natural connections between them as edges. Graph convolutional networks have been used to recognize actions from skeletal videos. We introduce a part-based graph convolutional network (PB-GCN) for this task, inspired by Deformable Part-based Models (DPMs). We divide the skeleton graph into four subgraphs with joints shared across them and learn a recognition model using a part-based graph convolutional network. We show that such a model improves performance of recognition, compared to a model using entire skeleton graph. Instead of using 3D joint coordinates as node features, we show that using relative coordinates and temporal displacements boosts performance. Our model achieves state-of-the-art performance on two challenging benchmark datasets NTURGB+D and HDM05, for skeletal action recognition.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)87.5PB-GCN
VideoNTU RGB+DAccuracy (CV)93.2PB-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)87.5PB-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)93.2PB-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CS)87.5PB-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CV)93.2PB-GCN
Activity RecognitionNTU RGB+DAccuracy (CS)87.5PB-GCN (Skeleton only)
Activity RecognitionNTU RGB+DAccuracy (CV)93.2PB-GCN (Skeleton only)
Activity RecognitionNTU RGB+DAccuracy (CS)87.5PB-GCN
Activity RecognitionNTU RGB+DAccuracy (CV)93.2PB-GCN
Action LocalizationNTU RGB+DAccuracy (CS)87.5PB-GCN
Action LocalizationNTU RGB+DAccuracy (CV)93.2PB-GCN
Action DetectionNTU RGB+DAccuracy (CS)87.5PB-GCN
Action DetectionNTU RGB+DAccuracy (CV)93.2PB-GCN
3D Action RecognitionNTU RGB+DAccuracy (CS)87.5PB-GCN
3D Action RecognitionNTU RGB+DAccuracy (CV)93.2PB-GCN
Action RecognitionNTU RGB+DAccuracy (CS)87.5PB-GCN (Skeleton only)
Action RecognitionNTU RGB+DAccuracy (CV)93.2PB-GCN (Skeleton only)
Action RecognitionNTU RGB+DAccuracy (CS)87.5PB-GCN
Action RecognitionNTU RGB+DAccuracy (CV)93.2PB-GCN

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22