TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Temporally-Adaptive Models for Efficient Video Understanding

Temporally-Adaptive Models for Efficient Video Understanding

Ziyuan Huang, Shiwei Zhang, Liang Pan, Zhiwu Qing, Yingya Zhang, Ziwei Liu, Marcelo H. Ang Jr

2023-08-10Action ClassificationVideo UnderstandingAction Recognition
PaperPDFCode(official)

Abstract

Spatial convolutions are extensively used in numerous deep video models. It fundamentally assumes spatio-temporal invariance, i.e., using shared weights for every location in different frames. This work presents Temporally-Adaptive Convolutions (TAdaConv) for video understanding, which shows that adaptive weight calibration along the temporal dimension is an efficient way to facilitate modeling complex temporal dynamics in videos. Specifically, TAdaConv empowers spatial convolutions with temporal modeling abilities by calibrating the convolution weights for each frame according to its local and global temporal context. Compared to existing operations for temporal modeling, TAdaConv is more efficient as it operates over the convolution kernels instead of the features, whose dimension is an order of magnitude smaller than the spatial resolutions. Further, kernel calibration brings an increased model capacity. Based on this readily plug-in operation TAdaConv as well as its extension, i.e., TAdaConvV2, we construct TAdaBlocks to empower ConvNeXt and Vision Transformer to have strong temporal modeling capabilities. Empirical results show TAdaConvNeXtV2 and TAdaFormer perform competitively against state-of-the-art convolutional and Transformer-based models in various video understanding benchmarks. Our codes and models are released at: https://github.com/alibaba-mmai-research/TAdaConv.

Results

TaskDatasetMetricValueModel
VideoKinetics-400Acc@189.9TAdaFormer-L/14
VideoKinetics-400Acc@186.4TAdaConvNeXtV2-B
Activity RecognitionEPIC-KITCHENS-100Action@151.8TAdaFormer-L/14
Activity RecognitionEPIC-KITCHENS-100Noun@164.1TAdaFormer-L/14
Activity RecognitionEPIC-KITCHENS-100Verb@171.7TAdaFormer-L/14
Activity RecognitionEPIC-KITCHENS-100Action@148.9TAdaConvNeXtV2-S
Activity RecognitionEPIC-KITCHENS-100Noun@160.2TAdaConvNeXtV2-S
Activity RecognitionEPIC-KITCHENS-100Verb@171TAdaConvNeXtV2-S
Activity RecognitionSomething-Something V1Top 1 Accuracy63.7TAdaFormer-L/14
Activity RecognitionSomething-Something V1Top 1 Accuracy60.7TAdaConvNeXtV2-B
Activity RecognitionSomething-Something V2Top-1 Accuracy73.6TAdaFormer-L/14
Activity RecognitionSomething-Something V2Top-1 Accuracy71.1TAdaConvNeXtV2-B
Action RecognitionEPIC-KITCHENS-100Action@151.8TAdaFormer-L/14
Action RecognitionEPIC-KITCHENS-100Noun@164.1TAdaFormer-L/14
Action RecognitionEPIC-KITCHENS-100Verb@171.7TAdaFormer-L/14
Action RecognitionEPIC-KITCHENS-100Action@148.9TAdaConvNeXtV2-S
Action RecognitionEPIC-KITCHENS-100Noun@160.2TAdaConvNeXtV2-S
Action RecognitionEPIC-KITCHENS-100Verb@171TAdaConvNeXtV2-S
Action RecognitionSomething-Something V1Top 1 Accuracy63.7TAdaFormer-L/14
Action RecognitionSomething-Something V1Top 1 Accuracy60.7TAdaConvNeXtV2-B
Action RecognitionSomething-Something V2Top-1 Accuracy73.6TAdaFormer-L/14
Action RecognitionSomething-Something V2Top-1 Accuracy71.1TAdaConvNeXtV2-B

Related Papers

VideoITG: Multimodal Video Understanding with Instructed Temporal Grounding2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17UGC-VideoCaptioner: An Omni UGC Video Detail Caption Model and New Benchmarks2025-07-15EmbRACE-3K: Embodied Reasoning and Action in Complex Environments2025-07-14Chat with AI: The Surprising Turn of Real-time Video Communication from Human to AI2025-07-14Beyond Appearance: Geometric Cues for Robust Video Instance Segmentation2025-07-08Omni-Video: Democratizing Unified Video Understanding and Generation2025-07-08MCAM: Multimodal Causal Analysis Model for Ego-Vehicle-Level Driving Video Understanding2025-07-08