TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Predictively Encoded Graph Convolutional Network for Noise...

Predictively Encoded Graph Convolutional Network for Noise-Robust Skeleton-based Action Recognition

Jongmin Yu, Yongsang Yoon, Moongu Jeon

2020-03-17Skeleton Based Action RecognitionAction Recognition
PaperPDFCode(official)

Abstract

In skeleton-based action recognition, graph convolutional networks (GCNs), which model human body skeletons using graphical components such as nodes and connections, have achieved remarkable performance recently. However, current state-of-the-art methods for skeleton-based action recognition usually work on the assumption that the completely observed skeletons will be provided. This may be problematic to apply this assumption in real scenarios since there is always a possibility that captured skeletons are incomplete or noisy. In this work, we propose a skeleton-based action recognition method which is robust to noise information of given skeleton features. The key insight of our approach is to train a model by maximizing the mutual information between normal and noisy skeletons using a predictive coding manner. We have conducted comprehensive experiments about skeleton-based action recognition with defected skeletons using NTU-RGB+D and Kinetics-Skeleton datasets. The experimental results demonstrate that our approach achieves outstanding performance when skeleton samples are noised compared with existing state-of-the-art methods.

Results

TaskDatasetMetricValueModel
VideoKinetics-Skeleton datasetAccuracy34.8PeGCN
VideoNTU RGB+DAccuracy (CS)85.6PeGCN
VideoNTU RGB+DAccuracy (CV)93.4PeGCN
Temporal Action LocalizationKinetics-Skeleton datasetAccuracy34.8PeGCN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)85.6PeGCN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)93.4PeGCN
Zero-Shot LearningKinetics-Skeleton datasetAccuracy34.8PeGCN
Zero-Shot LearningNTU RGB+DAccuracy (CS)85.6PeGCN
Zero-Shot LearningNTU RGB+DAccuracy (CV)93.4PeGCN
Activity RecognitionKinetics-Skeleton datasetAccuracy34.8PeGCN
Activity RecognitionNTU RGB+DAccuracy (CS)85.6PeGCN
Activity RecognitionNTU RGB+DAccuracy (CV)93.4PeGCN
Action LocalizationKinetics-Skeleton datasetAccuracy34.8PeGCN
Action LocalizationNTU RGB+DAccuracy (CS)85.6PeGCN
Action LocalizationNTU RGB+DAccuracy (CV)93.4PeGCN
Action DetectionKinetics-Skeleton datasetAccuracy34.8PeGCN
Action DetectionNTU RGB+DAccuracy (CS)85.6PeGCN
Action DetectionNTU RGB+DAccuracy (CV)93.4PeGCN
3D Action RecognitionKinetics-Skeleton datasetAccuracy34.8PeGCN
3D Action RecognitionNTU RGB+DAccuracy (CS)85.6PeGCN
3D Action RecognitionNTU RGB+DAccuracy (CV)93.4PeGCN
Action RecognitionKinetics-Skeleton datasetAccuracy34.8PeGCN
Action RecognitionNTU RGB+DAccuracy (CS)85.6PeGCN
Action RecognitionNTU RGB+DAccuracy (CV)93.4PeGCN

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22Active Multimodal Distillation for Few-shot Action Recognition2025-06-16