TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Relational Network for Skeleton-Based Action Recognition

Relational Network for Skeleton-Based Action Recognition

Wu Zheng, Lin Li, Zhao-Xiang Zhang, Yan Huang, Liang Wang

2018-05-07Skeleton Based Action RecognitionAction RecognitionTemporal Action Localization
PaperPDF

Abstract

With the fast development of effective and low-cost human skeleton capture systems, skeleton-based action recognition has attracted much attention recently. Most existing methods use Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) to extract spatio-temporal information embedded in the skeleton sequences for action recognition. However, these approaches are limited in the ability of relational modeling in a single skeleton, due to the loss of important structural information when converting the raw skeleton data to adapt to the input format of CNN or RNN. In this paper, we propose an Attentional Recurrent Relational Network-LSTM (ARRN-LSTM) to simultaneously model spatial configurations and temporal dynamics in skeletons for action recognition. We introduce the Recurrent Relational Network to learn the spatial features in a single skeleton, followed by a multi-layer LSTM to learn the temporal features in the skeleton sequences. Between the two modules, we design an adaptive attentional module to focus attention on the most discriminative parts in the single skeleton. To exploit the complementarity from different geometries in the skeleton for sufficient relational modeling, we design a two-stream architecture to learn the structural features among joints and lines simultaneously. Extensive experiments are conducted on several popular skeleton datasets and the results show that the proposed approach achieves better results than most mainstream methods.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
VideoNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
Temporal Action LocalizationNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
Temporal Action LocalizationNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
Zero-Shot LearningNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
Zero-Shot LearningNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
Activity RecognitionNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
Activity RecognitionNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
Action LocalizationNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
Action LocalizationNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
Action DetectionNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
Action DetectionNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
3D Action RecognitionNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
3D Action RecognitionNTU RGB+DAccuracy (CV)88.8ARRN-LSTM
Action RecognitionNTU RGB+DAccuracy (CS)80.7ARRN-LSTM
Action RecognitionNTU RGB+DAccuracy (CV)88.8ARRN-LSTM

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22