TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Actionlet-Dependent Contrastive Learning for Unsupervised ...

Actionlet-Dependent Contrastive Learning for Unsupervised Skeleton-Based Action Recognition

Lilang Lin, Jiahang Zhang, Jiaying Liu

2023-03-20CVPR 2023 1Unsupervised Skeleton Based Action RecognitionSelf-supervised Skeleton-based Action RecognitionSkeleton Based Action RecognitionContrastive LearningAction Recognition
PaperPDF

Abstract

The self-supervised pretraining paradigm has achieved great success in skeleton-based action recognition. However, these methods treat the motion and static parts equally, and lack an adaptive design for different parts, which has a negative impact on the accuracy of action recognition. To realize the adaptive action modeling of both parts, we propose an Actionlet-Dependent Contrastive Learning method (ActCLR). The actionlet, defined as the discriminative subset of the human skeleton, effectively decomposes motion regions for better action modeling. In detail, by contrasting with the static anchor without motion, we extract the motion region of the skeleton data, which serves as the actionlet, in an unsupervised manner. Then, centering on actionlet, a motion-adaptive data transformation method is built. Different data transformations are applied to actionlet and non-actionlet regions to introduce more diversity while maintaining their own characteristics. Meanwhile, we propose a semantic-aware feature pooling method to build feature representations among motion and static regions in a distinguished manner. Extensive experiments on NTU RGB+D and PKUMMD show that the proposed method achieves remarkable action recognition performance. More visualization and quantitative experiments demonstrate the effectiveness of our method. Our project website is available at https://langlandslin.github.io/projects/ActCLR/

Results

TaskDatasetMetricValueModel
VideoNTU RGB+DAccuracy (CS)84.33s-ActCLR
VideoNTU RGB+DAccuracy (CV)88.83s-ActCLR
Temporal Action LocalizationNTU RGB+DAccuracy (CS)84.33s-ActCLR
Temporal Action LocalizationNTU RGB+DAccuracy (CV)88.83s-ActCLR
Zero-Shot LearningNTU RGB+DAccuracy (CS)84.33s-ActCLR
Zero-Shot LearningNTU RGB+DAccuracy (CV)88.83s-ActCLR
Activity RecognitionNTU RGB+DAccuracy (CS)84.33s-ActCLR
Activity RecognitionNTU RGB+DAccuracy (CV)88.83s-ActCLR
Action LocalizationNTU RGB+DAccuracy (CS)84.33s-ActCLR
Action LocalizationNTU RGB+DAccuracy (CV)88.83s-ActCLR
Action DetectionNTU RGB+DAccuracy (CS)84.33s-ActCLR
Action DetectionNTU RGB+DAccuracy (CV)88.83s-ActCLR
3D Action RecognitionNTU RGB+DAccuracy (CS)84.33s-ActCLR
3D Action RecognitionNTU RGB+DAccuracy (CV)88.83s-ActCLR
Action RecognitionNTU RGB+DAccuracy (CS)84.33s-ActCLR
Action RecognitionNTU RGB+DAccuracy (CV)88.83s-ActCLR

Related Papers

SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16LLM-Driven Dual-Level Multi-Interest Modeling for Recommendation2025-07-15Latent Space Consistency for Sparse-View CT Reconstruction2025-07-15