TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/AutoGCN -- Towards Generic Human Activity Recognition with...

AutoGCN -- Towards Generic Human Activity Recognition with Neural Architecture Search

Felix Tempel, Inga Strümke, Espen Alexander F. Ihlen

2024-02-02Skeleton Based Action RecognitionHuman Activity RecognitionNeural Architecture SearchAction RecognitionActivity Recognition
PaperPDFCode(official)

Abstract

This paper introduces AutoGCN, a generic Neural Architecture Search (NAS) algorithm for Human Activity Recognition (HAR) using Graph Convolution Networks (GCNs). HAR has gained attention due to advances in deep learning, increased data availability, and enhanced computational capabilities. At the same time, GCNs have shown promising results in modeling relationships between body key points in a skeletal graph. While domain experts often craft dataset-specific GCN-based methods, their applicability beyond this specific context is severely limited. AutoGCN seeks to address this limitation by simultaneously searching for the ideal hyperparameters and architecture combination within a versatile search space using a reinforcement controller while balancing optimal exploration and exploitation behavior with a knowledge reservoir during the search process. We conduct extensive experiments on two large-scale datasets focused on skeleton-based action recognition to assess the proposed algorithm's performance. Our experimental results underscore the effectiveness of AutoGCN in constructing optimal GCN architectures for HAR, outperforming conventional NAS and GCN methods, as well as random search. These findings highlight the significance of a diverse search space and an expressive input representation to enhance the network performance and generalizability.

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)88.3AutoGCN
VideoNTU RGB+DAccuracy (CV)95.5AutoGCN
Temporal Action LocalizationNTU RGB+DAccuracy (CS)88.3AutoGCN
Temporal Action LocalizationNTU RGB+DAccuracy (CV)95.5AutoGCN
Zero-Shot LearningNTU RGB+DAccuracy (CS)88.3AutoGCN
Zero-Shot LearningNTU RGB+DAccuracy (CV)95.5AutoGCN
Activity RecognitionNTU RGB+DAccuracy (CS)88.3AutoGCN
Activity RecognitionNTU RGB+DAccuracy (CV)95.5AutoGCN
Action LocalizationNTU RGB+DAccuracy (CS)88.3AutoGCN
Action LocalizationNTU RGB+DAccuracy (CV)95.5AutoGCN
Action DetectionNTU RGB+DAccuracy (CS)88.3AutoGCN
Action DetectionNTU RGB+DAccuracy (CV)95.5AutoGCN
3D Action RecognitionNTU RGB+DAccuracy (CS)88.3AutoGCN
3D Action RecognitionNTU RGB+DAccuracy (CV)95.5AutoGCN
Action RecognitionNTU RGB+DAccuracy (CS)88.3AutoGCN
Action RecognitionNTU RGB+DAccuracy (CV)95.5AutoGCN

Related Papers

DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17ZKP-FedEval: Verifiable and Privacy-Preserving Federated Evaluation using Zero-Knowledge Proofs2025-07-15Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26SEZ-HARN: Self-Explainable Zero-shot Human Activity Recognition Network2025-06-25Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25