TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Richly Activated Graph Convolutional Network for Action Re...

Richly Activated Graph Convolutional Network for Action Recognition with Incomplete Skeletons

Yi-Fan Song, Zhang Zhang, Liang Wang

2019-05-16Skeleton Based Action RecognitionAction RecognitionTemporal Action Localization
PaperPDFCodeCodeCode

Abstract

Current methods for skeleton-based human action recognition usually work with completely observed skeletons. However, in real scenarios, it is prone to capture incomplete and noisy skeletons, which will deteriorate the performance of traditional models. To enhance the robustness of action recognition models to incomplete skeletons, we propose a multi-stream graph convolutional network (GCN) for exploring sufficient discriminative features distributed over all skeleton joints. Here, each stream of the network is only responsible for learning features from currently unactivated joints, which are distinguished by the class activation maps (CAM) obtained by preceding streams, so that the activated joints of the proposed method are obviously more than traditional methods. Thus, the proposed method is termed richly activated GCN (RA-GCN), where the richly discovered features will improve the robustness of the model. Compared to the state-of-the-art methods, the RA-GCN achieves comparable performance on the NTU RGB+D dataset. Moreover, on a synthetic occlusion dataset, the performance deterioration can be alleviated by the RA-GCN significantly.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)85.93s RA-GCN
VideoNTU RGB+DAccuracy (CV)93.53s RA-GCN
VideoNTU RGB+DAccuracy (CS)85.82s RA-GCN
VideoNTU RGB+DAccuracy (CV)932s RA-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)85.93s RA-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)93.53s RA-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)85.82s RA-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)932s RA-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CS)85.93s RA-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CV)93.53s RA-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CS)85.82s RA-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CV)932s RA-GCN
Activity RecognitionNTU RGB+DAccuracy (CS)85.93s RA-GCN
Activity RecognitionNTU RGB+DAccuracy (CV)93.53s RA-GCN
Activity RecognitionNTU RGB+DAccuracy (CS)85.82s RA-GCN
Activity RecognitionNTU RGB+DAccuracy (CV)932s RA-GCN
Action LocalizationNTU RGB+DAccuracy (CS)85.93s RA-GCN
Action LocalizationNTU RGB+DAccuracy (CV)93.53s RA-GCN
Action LocalizationNTU RGB+DAccuracy (CS)85.82s RA-GCN
Action LocalizationNTU RGB+DAccuracy (CV)932s RA-GCN
Action DetectionNTU RGB+DAccuracy (CS)85.93s RA-GCN
Action DetectionNTU RGB+DAccuracy (CV)93.53s RA-GCN
Action DetectionNTU RGB+DAccuracy (CS)85.82s RA-GCN
Action DetectionNTU RGB+DAccuracy (CV)932s RA-GCN
3D Action RecognitionNTU RGB+DAccuracy (CS)85.93s RA-GCN
3D Action RecognitionNTU RGB+DAccuracy (CV)93.53s RA-GCN
3D Action RecognitionNTU RGB+DAccuracy (CS)85.82s RA-GCN
3D Action RecognitionNTU RGB+DAccuracy (CV)932s RA-GCN
Action RecognitionNTU RGB+DAccuracy (CS)85.93s RA-GCN
Action RecognitionNTU RGB+DAccuracy (CV)93.53s RA-GCN
Action RecognitionNTU RGB+DAccuracy (CS)85.82s RA-GCN
Action RecognitionNTU RGB+DAccuracy (CV)932s RA-GCN

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22