TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Richly Activated Graph Convolutional Network for Robust Sk...

Richly Activated Graph Convolutional Network for Robust Skeleton-based Action Recognition

Yi-Fan Song, Zhang Zhang, Caifeng Shan, Liang Wang

2020-08-09Skeleton Based Action RecognitionAction RecognitionTemporal Action Localization
PaperPDFCode(official)CodeCode

Abstract

Current methods for skeleton-based human action recognition usually work with complete skeletons. However, in real scenarios, it is inevitable to capture incomplete or noisy skeletons, which could significantly deteriorate the performance of current methods when some informative joints are occluded or disturbed. To improve the robustness of action recognition models, a multi-stream graph convolutional network (GCN) is proposed to explore sufficient discriminative features spreading over all skeleton joints, so that the distributed redundant representation reduces the sensitivity of the action models to non-standard skeletons. Concretely, the backbone GCN is extended by a series of ordered streams which is responsible for learning discriminative features from the joints less activated by preceding streams. Here, the activation degrees of skeleton joints of each GCN stream are measured by the class activation maps (CAM), and only the information from the unactivated joints will be passed to the next stream, by which rich features over all active joints are obtained. Thus, the proposed method is termed richly activated GCN (RA-GCN). Compared to the state-of-the-art (SOTA) methods, the RA-GCN achieves comparable performance on the standard NTU RGB+D 60 and 120 datasets. More crucially, on the synthetic occlusion and jittering datasets, the performance deterioration due to the occluded and disturbed joints can be significantly alleviated by utilizing the proposed RA-GCN.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)87.33s RA-GCN
VideoNTU RGB+DAccuracy (CV)93.63s RA-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)87.33s RA-GCN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)93.63s RA-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CS)87.33s RA-GCN
Zero-Shot LearningNTU RGB+DAccuracy (CV)93.63s RA-GCN
Activity RecognitionNTU RGB+DAccuracy (CS)87.33s RA-GCN
Activity RecognitionNTU RGB+DAccuracy (CV)93.63s RA-GCN
Action LocalizationNTU RGB+DAccuracy (CS)87.33s RA-GCN
Action LocalizationNTU RGB+DAccuracy (CV)93.63s RA-GCN
Action DetectionNTU RGB+DAccuracy (CS)87.33s RA-GCN
Action DetectionNTU RGB+DAccuracy (CV)93.63s RA-GCN
3D Action RecognitionNTU RGB+DAccuracy (CS)87.33s RA-GCN
3D Action RecognitionNTU RGB+DAccuracy (CV)93.63s RA-GCN
Action RecognitionNTU RGB+DAccuracy (CS)87.33s RA-GCN
Action RecognitionNTU RGB+DAccuracy (CV)93.63s RA-GCN

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22