TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Multiview Transformers for Video Recognition

Multiview Transformers for Video Recognition

Shen Yan, Xuehan Xiong, Anurag Arnab, Zhichao Lu, Mi Zhang, Chen Sun, Cordelia Schmid

2022-01-12CVPR 2022 1Action ClassificationVideo UnderstandingAction Recognition
PaperPDFCode(official)

Abstract

Video understanding requires reasoning at multiple spatiotemporal resolutions -- from short fine-grained motions to events taking place over longer durations. Although transformer architectures have recently advanced the state-of-the-art, they have not explicitly modelled different spatiotemporal resolutions. To this end, we present Multiview Transformers for Video Recognition (MTV). Our model consists of separate encoders to represent different views of the input video with lateral connections to fuse information across views. We present thorough ablation studies of our model and show that MTV consistently performs better than single-view counterparts in terms of accuracy and computational cost across a range of model sizes. Furthermore, we achieve state-of-the-art results on six standard datasets, and improve even further with large-scale pretraining. Code and checkpoints are available at: https://github.com/google-research/scenic/tree/main/scenic/projects/mtv.

Results

TaskDatasetMetricValueModel
VideoKinetics-700Top-1 Accuracy83.4MTV-H (WTS 60M)
VideoKinetics-700Top-5 Accuracy96.2MTV-H (WTS 60M)
VideoMiTTop 1 Accuracy47.2MTV-H (WTS 60M)
VideoMiTTop 5 Accuracy75.7MTV-H (WTS 60M)
VideoKinetics-400Acc@189.9MTV-H (WTS 60M)
VideoKinetics-400Acc@598.3MTV-H (WTS 60M)
VideoKinetics-600Top-1 Accuracy90.3MTV-H (WTS 60M)
VideoKinetics-600Top-5 Accuracy98.5MTV-H (WTS 60M)
Activity RecognitionEPIC-KITCHENS-100Action@150.5MTV-B (WTS 60M)
Activity RecognitionEPIC-KITCHENS-100Noun@163.9MTV-B (WTS 60M)
Activity RecognitionEPIC-KITCHENS-100Verb@169.9MTV-B (WTS 60M)
Activity RecognitionSomething-Something V2Top-1 Accuracy68.5MTV-B
Activity RecognitionSomething-Something V2Top-5 Accuracy90.4MTV-B
Action RecognitionEPIC-KITCHENS-100Action@150.5MTV-B (WTS 60M)
Action RecognitionEPIC-KITCHENS-100Noun@163.9MTV-B (WTS 60M)
Action RecognitionEPIC-KITCHENS-100Verb@169.9MTV-B (WTS 60M)
Action RecognitionSomething-Something V2Top-1 Accuracy68.5MTV-B
Action RecognitionSomething-Something V2Top-5 Accuracy90.4MTV-B

Related Papers

VideoITG: Multimodal Video Understanding with Instructed Temporal Grounding2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17UGC-VideoCaptioner: An Omni UGC Video Detail Caption Model and New Benchmarks2025-07-15EmbRACE-3K: Embodied Reasoning and Action in Complex Environments2025-07-14Chat with AI: The Surprising Turn of Real-time Video Communication from Human to AI2025-07-14Beyond Appearance: Geometric Cues for Robust Video Instance Segmentation2025-07-08Omni-Video: Democratizing Unified Video Understanding and Generation2025-07-08MCAM: Multimodal Causal Analysis Model for Ego-Vehicle-Level Driving Video Understanding2025-07-08