TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/View Adaptive Recurrent Neural Networks for High Performan...

View Adaptive Recurrent Neural Networks for High Performance Human Action Recognition from Skeleton Data

Pengfei Zhang, Cuiling Lan, Junliang Xing, Wen-Jun Zeng, Jianru Xue, Nanning Zheng

2017-03-24ICCV 2017 10Skeleton Based Action RecognitionAction RecognitionTemporal Action Localization
PaperPDFCode(official)

Abstract

Skeleton-based human action recognition has recently attracted increasing attention due to the popularity of 3D skeleton data. One main challenge lies in the large view variations in captured human actions. We propose a novel view adaptation scheme to automatically regulate observation viewpoints during the occurrence of an action. Rather than re-positioning the skeletons based on a human defined prior criterion, we design a view adaptive recurrent neural network (RNN) with LSTM architecture, which enables the network itself to adapt to the most suitable observation viewpoints from end to end. Extensive experiment analyses show that the proposed view adaptive RNN model strives to (1) transform the skeletons of various views to much more consistent viewpoints and (2) maintain the continuity of the action rather than transforming every frame to the same position with the same body orientation. Our model achieves significant improvement over the state-of-the-art approaches on three benchmark datasets.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)79.2VA-LSTM
VideoNTU RGB+DAccuracy (CV)87.6VA-LSTM
Temporal Action LocalizationNTU RGB+DAccuracy (CS)79.2VA-LSTM
Temporal Action LocalizationNTU RGB+DAccuracy (CV)87.6VA-LSTM
Zero-Shot LearningNTU RGB+DAccuracy (CS)79.2VA-LSTM
Zero-Shot LearningNTU RGB+DAccuracy (CV)87.6VA-LSTM
Activity RecognitionNTU RGB+DAccuracy (CS)79.2VA-LSTM
Activity RecognitionNTU RGB+DAccuracy (CV)87.6VA-LSTM
Action LocalizationNTU RGB+DAccuracy (CS)79.2VA-LSTM
Action LocalizationNTU RGB+DAccuracy (CV)87.6VA-LSTM
Action DetectionNTU RGB+DAccuracy (CS)79.2VA-LSTM
Action DetectionNTU RGB+DAccuracy (CV)87.6VA-LSTM
3D Action RecognitionNTU RGB+DAccuracy (CS)79.2VA-LSTM
3D Action RecognitionNTU RGB+DAccuracy (CV)87.6VA-LSTM
Action RecognitionNTU RGB+DAccuracy (CS)79.2VA-LSTM
Action RecognitionNTU RGB+DAccuracy (CV)87.6VA-LSTM

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22