TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Where a Strong Backbone Meets Strong Features -- ActionFor...

Where a Strong Backbone Meets Strong Features -- ActionFormer for Ego4D Moment Queries Challenge

Fangzhou Mu, Sicheng Mo, Gillian Wang, Yin Li

2022-11-16Action LocalizationMoment QueriesTemporal Action Localization
PaperPDFCode(official)Code(official)

Abstract

This report describes our submission to the Ego4D Moment Queries Challenge 2022. Our submission builds on ActionFormer, the state-of-the-art backbone for temporal action localization, and a trio of strong video features from SlowFast, Omnivore and EgoVLP. Our solution is ranked 2nd on the public leaderboard with 21.76% average mAP on the test set, which is nearly three times higher than the official baseline. Further, we obtain 42.54% Recall@1x at tIoU=0.5 on the test set, outperforming the top-ranked solution by a significant margin of 1.41 absolute percentage points. Our code is available at https://github.com/happyharrycn/actionformer_release.

Results

TaskDatasetMetricValueModel
VideoEgo4D MQ testAverage mAP21.76ActionFormer (SlowFast+Omnivore+EgoVLP)
VideoEgo4D MQ testRecall@1x (tIoU=0.5)42.54ActionFormer (SlowFast+Omnivore+EgoVLP)
VideoEgo4D MQ valAverage mAP21.4ActionFormer (SlowFast+Omnivore+EgoVLP)
VideoEgo4D MQ valRecall@1x (tIoU=0.5)38.73ActionFormer (SlowFast+Omnivore+EgoVLP)
Temporal Action LocalizationEgo4D MQ testAverage mAP21.76ActionFormer (SlowFast+Omnivore+EgoVLP)
Temporal Action LocalizationEgo4D MQ testRecall@1x (tIoU=0.5)42.54ActionFormer (SlowFast+Omnivore+EgoVLP)
Temporal Action LocalizationEgo4D MQ valAverage mAP21.4ActionFormer (SlowFast+Omnivore+EgoVLP)
Temporal Action LocalizationEgo4D MQ valRecall@1x (tIoU=0.5)38.73ActionFormer (SlowFast+Omnivore+EgoVLP)
Zero-Shot LearningEgo4D MQ testAverage mAP21.76ActionFormer (SlowFast+Omnivore+EgoVLP)
Zero-Shot LearningEgo4D MQ testRecall@1x (tIoU=0.5)42.54ActionFormer (SlowFast+Omnivore+EgoVLP)
Zero-Shot LearningEgo4D MQ valAverage mAP21.4ActionFormer (SlowFast+Omnivore+EgoVLP)
Zero-Shot LearningEgo4D MQ valRecall@1x (tIoU=0.5)38.73ActionFormer (SlowFast+Omnivore+EgoVLP)
Action LocalizationEgo4D MQ testAverage mAP21.76ActionFormer (SlowFast+Omnivore+EgoVLP)
Action LocalizationEgo4D MQ testRecall@1x (tIoU=0.5)42.54ActionFormer (SlowFast+Omnivore+EgoVLP)
Action LocalizationEgo4D MQ valAverage mAP21.4ActionFormer (SlowFast+Omnivore+EgoVLP)
Action LocalizationEgo4D MQ valRecall@1x (tIoU=0.5)38.73ActionFormer (SlowFast+Omnivore+EgoVLP)

Related Papers

DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Zero-Shot Temporal Interaction Localization for Egocentric Videos2025-06-04OSGNet @ Ego4D Episodic Memory Challenge 20252025-06-04A Review on Coarse to Fine-Grained Animal Action Recognition2025-06-01LLM-powered Query Expansion for Enhancing Boundary Prediction in Language-driven Action Localization2025-05-30CLIP-AE: CLIP-assisted Cross-view Audio-Visual Enhancement for Unsupervised Temporal Action Localization2025-05-29DeepConvContext: A Multi-Scale Approach to Timeseries Classification in Human Activity Recognition2025-05-27