TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Video Transformer Network

Video Transformer Network

Daniel Neimark, Omri Bar, Maya Zohar, Dotan Asselmann

2021-02-01Action ClassificationVideo RecognitionAction RecognitionAction Recognition In Videos
PaperPDFCode(official)

Abstract

This paper presents VTN, a transformer-based framework for video recognition. Inspired by recent developments in vision transformers, we ditch the standard approach in video action recognition that relies on 3D ConvNets and introduce a method that classifies actions by attending to the entire video sequence information. Our approach is generic and builds on top of any given 2D spatial network. In terms of wall runtime, it trains $16.1\times$ faster and runs $5.1\times$ faster during inference while maintaining competitive accuracy compared to other state-of-the-art methods. It enables whole video analysis, via a single end-to-end pass, while requiring $1.5\times$ fewer GFLOPs. We report competitive results on Kinetics-400 and present an ablation study of VTN properties and the trade-off between accuracy and inference speed. We hope our approach will serve as a new baseline and start a fresh line of research in the video recognition domain. Code and models are available at: https://github.com/bomri/SlowFast/blob/master/projects/vtn/README.md

Results

TaskDatasetMetricValueModel
VideoMiTTop 1 Accuracy37.4VTN
VideoMiTTop 5 Accuracy65.4VTN
VideoKinetics-400Acc@179.8ViT-B-VTN+ ImageNet-21K (84.0 [10])
VideoKinetics-400Acc@178.6ViT-B-VTN (3 layers, ImageNet pretrain)
VideoKinetics-400Acc@593.7ViT-B-VTN (3 layers, ImageNet pretrain)
VideoKinetics-400Acc@594.2ViT-B-VTN+ ImageNet-21K (84.0 [10])
VideoKinetics-400Acc@593.4ViT-B-VTN (1 layer, ImageNet pretrain)

Related Papers

A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Zero-shot Skeleton-based Action Recognition with Prototype-guided Feature Alignment2025-07-01EgoAdapt: Adaptive Multisensory Distillation and Policy Learning for Efficient Egocentric Perception2025-06-26Feature Hallucination for Self-supervised Action Recognition2025-06-25CARMA: Context-Aware Situational Grounding of Human-Robot Group Interactions by Combining Vision-Language Models with Object and Action Recognition2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Adapting Vision-Language Models for Evaluating World Models2025-06-22