TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Boundary Content Graph Neural Network for Temporal Action ...

Boundary Content Graph Neural Network for Temporal Action Proposal Generation

Yueran Bai, Yingying Wang, Yunhai Tong, Yang Yang, Qiyue Liu, Junhui Liu

2020-08-04ECCV 2020 8Action DetectionTemporal Action Proposal GenerationTemporal Action LocalizationAction Understanding
PaperPDF

Abstract

Temporal action proposal generation plays an important role in video action understanding, which requires localizing high-quality action content precisely. However, generating temporal proposals with both precise boundaries and high-quality action content is extremely challenging. To address this issue, we propose a novel Boundary Content Graph Neural Network (BC-GNN) to model the insightful relations between the boundary and action content of temporal proposals by the graph neural networks. In BC-GNN, the boundaries and content of temporal proposals are taken as the nodes and edges of the graph neural network, respectively, where they are spontaneously linked. Then a novel graph computation operation is proposed to update features of edges and nodes. After that, one updated edge and two nodes it connects are used to predict boundary probabilities and content confidence score, which will be combined to generate a final high-quality proposal. Experiments are conducted on two mainstream datasets: ActivityNet-1.3 and THUMOS14. Without the bells and whistles, BC-GNN outperforms previous state-of-the-art methods in both temporal action proposal and temporal action detection tasks.

Results

TaskDatasetMetricValueModel
VideoActivityNet-1.3mAP34.26BC-GNN
VideoActivityNet-1.3mAP IOU@0.550.56BC-GNN
VideoActivityNet-1.3mAP IOU@0.7534.75BC-GNN
VideoActivityNet-1.3mAP IOU@0.959.37BC-GNN
Temporal Action LocalizationActivityNet-1.3mAP34.26BC-GNN
Temporal Action LocalizationActivityNet-1.3mAP IOU@0.550.56BC-GNN
Temporal Action LocalizationActivityNet-1.3mAP IOU@0.7534.75BC-GNN
Temporal Action LocalizationActivityNet-1.3mAP IOU@0.959.37BC-GNN
Zero-Shot LearningActivityNet-1.3mAP34.26BC-GNN
Zero-Shot LearningActivityNet-1.3mAP IOU@0.550.56BC-GNN
Zero-Shot LearningActivityNet-1.3mAP IOU@0.7534.75BC-GNN
Zero-Shot LearningActivityNet-1.3mAP IOU@0.959.37BC-GNN
Action LocalizationActivityNet-1.3mAP34.26BC-GNN
Action LocalizationActivityNet-1.3mAP IOU@0.550.56BC-GNN
Action LocalizationActivityNet-1.3mAP IOU@0.7534.75BC-GNN
Action LocalizationActivityNet-1.3mAP IOU@0.959.37BC-GNN

Related Papers

DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16LLaVA-Pose: Enhancing Human Pose and Action Understanding via Keypoint-Integrated Instruction Tuning2025-06-26CBF-AFA: Chunk-Based Multi-SSL Fusion for Automatic Fluency Assessment2025-06-25MultiHuman-Testbench: Benchmarking Image Generation for Multiple Humans2025-06-25Including Semantic Information via Word Embeddings for Skeleton-based Action Recognition2025-06-23Distributed Activity Detection for Cell-Free Hybrid Near-Far Field Communications2025-06-17Zero-Shot Temporal Interaction Localization for Egocentric Videos2025-06-04Speaker Diarization with Overlapping Community Detection Using Graph Attention Networks and Label Propagation Algorithm2025-06-03