TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Self-Similarity in Space and Time as Generalized ...

Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition

Heeseung Kwon, Manjin Kim, Suha Kwak, Minsu Cho

2021-02-14ICCV 2021 10Video UnderstandingAction RecognitionTemporal Action Localization
PaperPDFCode(official)

Abstract

Spatio-temporal convolution often fails to learn motion dynamics in videos and thus an effective motion representation is required for video understanding in the wild. In this paper, we propose a rich and robust motion representation based on spatio-temporal self-similarity (STSS). Given a sequence of frames, STSS represents each local region as similarities to its neighbors in space and time. By converting appearance features into relational values, it enables the learner to better recognize structural patterns in space and time. We leverage the whole volume of STSS and let our model learn to extract an effective motion representation from it. The proposed neural block, dubbed SELFY, can be easily inserted into neural architectures and trained end-to-end without additional supervision. With a sufficient volume of the neighborhood in space and time, it effectively captures long-term interaction and fast motion in the video, leading to robust action recognition. Our experimental analysis demonstrates its superiority over previous methods for motion modeling as well as its complementarity to spatio-temporal features from direct convolution. On the standard action recognition benchmarks, Something-Something-V1 & V2, Diving-48, and FineGym, the proposed method achieves the state-of-the-art results.

Results

TaskDatasetMetricValueModel
Activity RecognitionSomething-Something V1Top 1 Accuracy56.6SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Activity RecognitionSomething-Something V1Top 5 Accuracy84.4SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Activity RecognitionSomething-Something V1Top 1 Accuracy55.8SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Activity RecognitionSomething-Something V1Top 5 Accuracy83.9SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Activity RecognitionSomething-Something V1Top 1 Accuracy54.3SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Activity RecognitionSomething-Something V1Top 5 Accuracy82.9SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Activity RecognitionSomething-Something V2Top-1 Accuracy67.7SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Activity RecognitionSomething-Something V2Top-5 Accuracy91.1SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Activity RecognitionSomething-Something V2Top-1 Accuracy67.4SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Activity RecognitionSomething-Something V2Top-5 Accuracy91SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Activity RecognitionSomething-Something V2Top-1 Accuracy65.7SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Activity RecognitionSomething-Something V2Top-5 Accuracy89.8SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Action RecognitionSomething-Something V1Top 1 Accuracy56.6SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Action RecognitionSomething-Something V1Top 5 Accuracy84.4SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Action RecognitionSomething-Something V1Top 1 Accuracy55.8SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Action RecognitionSomething-Something V1Top 5 Accuracy83.9SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Action RecognitionSomething-Something V1Top 1 Accuracy54.3SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Action RecognitionSomething-Something V1Top 5 Accuracy82.9SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Action RecognitionSomething-Something V2Top-1 Accuracy67.7SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Action RecognitionSomething-Something V2Top-5 Accuracy91.1SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, 2 clips)
Action RecognitionSomething-Something V2Top-1 Accuracy67.4SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Action RecognitionSomething-Something V2Top-5 Accuracy91SELFYNet-TSM-R50En (8+16 frames, ImageNet pretrained, a single clip)
Action RecognitionSomething-Something V2Top-1 Accuracy65.7SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)
Action RecognitionSomething-Something V2Top-5 Accuracy89.8SELFYNet-TSM-R50 (16 frames, ImageNet pretrained)

Related Papers

VideoITG: Multimodal Video Understanding with Instructed Temporal Grounding2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16UGC-VideoCaptioner: An Omni UGC Video Detail Caption Model and New Benchmarks2025-07-15EmbRACE-3K: Embodied Reasoning and Action in Complex Environments2025-07-14Chat with AI: The Surprising Turn of Real-time Video Communication from Human to AI2025-07-14Beyond Appearance: Geometric Cues for Robust Video Instance Segmentation2025-07-08Omni-Video: Democratizing Unified Video Understanding and Generation2025-07-08