TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Activity Graph Transformer for Temporal Action Localization

Activity Graph Transformer for Temporal Action Localization

Megha Nawhal, Greg Mori

2021-01-21Action LocalizationTemporal Action Localization
PaperPDF

Abstract

We introduce Activity Graph Transformer, an end-to-end learnable model for temporal action localization, that receives a video as input and directly predicts a set of action instances that appear in the video. Detecting and localizing action instances in untrimmed videos requires reasoning over multiple action instances in a video. The dominant paradigms in the literature process videos temporally to either propose action regions or directly produce frame-level detections. However, sequential processing of videos is problematic when the action instances have non-sequential dependencies and/or non-linear temporal ordering, such as overlapping action instances or re-occurrence of action instances over the course of the video. In this work, we capture this non-linear temporal structure by reasoning over the videos as non-sequential entities in the form of graphs. We evaluate our model on challenging datasets: THUMOS14, Charades, and EPIC-Kitchens-100. Our results show that our proposed model outperforms the state-of-the-art by a considerable margin.

Results

TaskDatasetMetricValueModel
VideoTHUMOS’14mAP IOU@0.172.1AGT (Ours)
VideoTHUMOS’14mAP IOU@0.269.8AGT (Ours)
VideoTHUMOS’14mAP IOU@0.365AGT (Ours)
VideoTHUMOS’14mAP IOU@0.458.1AGT (Ours)
VideoTHUMOS’14mAP IOU@0.550.2AGT (Ours)
Temporal Action LocalizationTHUMOS’14mAP IOU@0.172.1AGT (Ours)
Temporal Action LocalizationTHUMOS’14mAP IOU@0.269.8AGT (Ours)
Temporal Action LocalizationTHUMOS’14mAP IOU@0.365AGT (Ours)
Temporal Action LocalizationTHUMOS’14mAP IOU@0.458.1AGT (Ours)
Temporal Action LocalizationTHUMOS’14mAP IOU@0.550.2AGT (Ours)
Zero-Shot LearningTHUMOS’14mAP IOU@0.172.1AGT (Ours)
Zero-Shot LearningTHUMOS’14mAP IOU@0.269.8AGT (Ours)
Zero-Shot LearningTHUMOS’14mAP IOU@0.365AGT (Ours)
Zero-Shot LearningTHUMOS’14mAP IOU@0.458.1AGT (Ours)
Zero-Shot LearningTHUMOS’14mAP IOU@0.550.2AGT (Ours)
Action LocalizationTHUMOS’14mAP IOU@0.172.1AGT (Ours)
Action LocalizationTHUMOS’14mAP IOU@0.269.8AGT (Ours)
Action LocalizationTHUMOS’14mAP IOU@0.365AGT (Ours)
Action LocalizationTHUMOS’14mAP IOU@0.458.1AGT (Ours)
Action LocalizationTHUMOS’14mAP IOU@0.550.2AGT (Ours)

Related Papers

DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Zero-Shot Temporal Interaction Localization for Egocentric Videos2025-06-04A Review on Coarse to Fine-Grained Animal Action Recognition2025-06-01LLM-powered Query Expansion for Enhancing Boundary Prediction in Language-driven Action Localization2025-05-30CLIP-AE: CLIP-assisted Cross-view Audio-Visual Enhancement for Unsupervised Temporal Action Localization2025-05-29DeepConvContext: A Multi-Scale Approach to Timeseries Classification in Human Activity Recognition2025-05-27ProTAL: A Drag-and-Link Video Programming Framework for Temporal Action Localization2025-05-23