TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Adversarial Background-Aware Loss for Weakly-supervised Te...

Adversarial Background-Aware Loss for Weakly-supervised Temporal Activity Localization

Kyle Min, Jason J. Corso

2020-07-13ECCV 2020 8Weakly Supervised Action LocalizationMetric LearningWeakly-supervised Temporal Action Localization
PaperPDFCode(official)

Abstract

Temporally localizing activities within untrimmed videos has been extensively studied in recent years. Despite recent advances, existing methods for weakly-supervised temporal activity localization struggle to recognize when an activity is not occurring. To address this issue, we propose a novel method named A2CL-PT. Two triplets of the feature space are considered in our approach: one triplet is used to learn discriminative features for each activity class, and the other one is used to distinguish the features where no activity occurs (i.e. background features) from activity-related features for each video. To further improve the performance, we build our network using two parallel branches which operate in an adversarial way: the first branch localizes the most salient activities of a video and the second one finds other supplementary activities from non-localized parts of the video. Extensive experiments performed on THUMOS14 and ActivityNet datasets demonstrate that our proposed method is effective. Specifically, the average mAP of IoU thresholds from 0.1 to 0.9 on the THUMOS14 dataset is significantly improved from 27.9% to 30.0%.

Results

TaskDatasetMetricValueModel
VideoTHUMOS’14mAP IOU@0.161.2A2CL-PT
VideoTHUMOS’14mAP IOU@0.256.1A2CL-PT
VideoTHUMOS’14mAP IOU@0.348.1A2CL-PT
VideoTHUMOS’14mAP IOU@0.439A2CL-PT
VideoTHUMOS’14mAP IOU@0.530.1A2CL-PT
VideoTHUMOS’14mAP IOU@0.619.2A2CL-PT
VideoTHUMOS’14mAP IOU@0.710.6A2CL-PT
VideoTHUMOS’14mAP IOU@0.84.8A2CL-PT
VideoTHUMOS’14mAP IOU@0.91A2CL-PT
VideoTHUMOS’14mAP@AVG(0.1:0.9)30A2CL-PT
VideoTHUMOS 2014mAP@0.1:0.546.9A2CL-PT
VideoTHUMOS 2014mAP@0.1:0.737.8A2CL-PT
VideoTHUMOS 2014mAP@0.530.1A2CL-PT
VideoTHUMOS14avg-mAP (0.1-0.5)46.9A2CL-PT
VideoTHUMOS14avg-mAP (0.1:0.7)37.8A2CL-PT
VideoTHUMOS14avg-mAP (0.3-0.7)30.6A2CL-PT
VideoTHUMOS’14mAP@0.530.1A2CL-PT
VideoActivityNet-1.3mAP@0.536.8A2CL-PT
VideoActivityNet-1.3mAP@0.5:0.9522.5A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.161.2A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.256.1A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.348.1A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.439A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.530.1A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.619.2A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.710.6A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.84.8A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP IOU@0.91A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP@AVG(0.1:0.9)30A2CL-PT
Temporal Action LocalizationTHUMOS 2014mAP@0.1:0.546.9A2CL-PT
Temporal Action LocalizationTHUMOS 2014mAP@0.1:0.737.8A2CL-PT
Temporal Action LocalizationTHUMOS 2014mAP@0.530.1A2CL-PT
Temporal Action LocalizationTHUMOS14avg-mAP (0.1-0.5)46.9A2CL-PT
Temporal Action LocalizationTHUMOS14avg-mAP (0.1:0.7)37.8A2CL-PT
Temporal Action LocalizationTHUMOS14avg-mAP (0.3-0.7)30.6A2CL-PT
Temporal Action LocalizationTHUMOS’14mAP@0.530.1A2CL-PT
Temporal Action LocalizationActivityNet-1.3mAP@0.536.8A2CL-PT
Temporal Action LocalizationActivityNet-1.3mAP@0.5:0.9522.5A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.161.2A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.256.1A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.348.1A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.439A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.530.1A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.619.2A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.710.6A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.84.8A2CL-PT
Zero-Shot LearningTHUMOS’14mAP IOU@0.91A2CL-PT
Zero-Shot LearningTHUMOS’14mAP@AVG(0.1:0.9)30A2CL-PT
Zero-Shot LearningTHUMOS 2014mAP@0.1:0.546.9A2CL-PT
Zero-Shot LearningTHUMOS 2014mAP@0.1:0.737.8A2CL-PT
Zero-Shot LearningTHUMOS 2014mAP@0.530.1A2CL-PT
Zero-Shot LearningTHUMOS14avg-mAP (0.1-0.5)46.9A2CL-PT
Zero-Shot LearningTHUMOS14avg-mAP (0.1:0.7)37.8A2CL-PT
Zero-Shot LearningTHUMOS14avg-mAP (0.3-0.7)30.6A2CL-PT
Zero-Shot LearningTHUMOS’14mAP@0.530.1A2CL-PT
Zero-Shot LearningActivityNet-1.3mAP@0.536.8A2CL-PT
Zero-Shot LearningActivityNet-1.3mAP@0.5:0.9522.5A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.161.2A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.256.1A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.348.1A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.439A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.530.1A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.619.2A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.710.6A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.84.8A2CL-PT
Action LocalizationTHUMOS’14mAP IOU@0.91A2CL-PT
Action LocalizationTHUMOS’14mAP@AVG(0.1:0.9)30A2CL-PT
Action LocalizationTHUMOS 2014mAP@0.1:0.546.9A2CL-PT
Action LocalizationTHUMOS 2014mAP@0.1:0.737.8A2CL-PT
Action LocalizationTHUMOS 2014mAP@0.530.1A2CL-PT
Action LocalizationTHUMOS14avg-mAP (0.1-0.5)46.9A2CL-PT
Action LocalizationTHUMOS14avg-mAP (0.1:0.7)37.8A2CL-PT
Action LocalizationTHUMOS14avg-mAP (0.3-0.7)30.6A2CL-PT
Action LocalizationTHUMOS’14mAP@0.530.1A2CL-PT
Action LocalizationActivityNet-1.3mAP@0.536.8A2CL-PT
Action LocalizationActivityNet-1.3mAP@0.5:0.9522.5A2CL-PT
Weakly Supervised Action LocalizationTHUMOS 2014mAP@0.1:0.546.9A2CL-PT
Weakly Supervised Action LocalizationTHUMOS 2014mAP@0.1:0.737.8A2CL-PT
Weakly Supervised Action LocalizationTHUMOS 2014mAP@0.530.1A2CL-PT
Weakly Supervised Action LocalizationTHUMOS14avg-mAP (0.1-0.5)46.9A2CL-PT
Weakly Supervised Action LocalizationTHUMOS14avg-mAP (0.1:0.7)37.8A2CL-PT
Weakly Supervised Action LocalizationTHUMOS14avg-mAP (0.3-0.7)30.6A2CL-PT
Weakly Supervised Action LocalizationTHUMOS’14mAP@0.530.1A2CL-PT
Weakly Supervised Action LocalizationActivityNet-1.3mAP@0.536.8A2CL-PT
Weakly Supervised Action LocalizationActivityNet-1.3mAP@0.5:0.9522.5A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.161.2A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.256.1A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.348.1A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.439A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.530.1A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.619.2A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.710.6A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.84.8A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP IOU@0.91A2CL-PT
Weakly-supervised Temporal Action LocalizationTHUMOS’14mAP@AVG(0.1:0.9)30A2CL-PT

Related Papers

Unsupervised Ground Metric Learning2025-07-17Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16$\texttt{Droid}$: A Resource Suite for AI-Generated Code Detection2025-07-11Grid-Reg: Grid-Based SAR and Optical Image Registration Across Platforms2025-07-06Dare to Plagiarize? Plagiarized Painting Recognition and Retrieval2025-06-29Multimodal Information Retrieval for Open World with Edit Distance Weak Supervision2025-06-25Offline Goal-Conditioned Reinforcement Learning with Projective Quasimetric Planning2025-06-23AbRank: A Benchmark Dataset and Metric-Learning Framework for Antibody-Antigen Affinity Ranking2025-06-21