TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/D3G: Exploring Gaussian Prior for Temporal Sentence Ground...

D3G: Exploring Gaussian Prior for Temporal Sentence Grounding with Glance Annotation

Hanjun Li, Xiujun Shu, Sunan He, Ruizhi Qiao, Wei Wen, Taian Guo, Bei Gan, Xing Sun

2023-08-08ICCV 2023 1Temporal Sentence GroundingContrastive Learning
PaperPDFCode(official)

Abstract

Temporal sentence grounding (TSG) aims to locate a specific moment from an untrimmed video with a given natural language query. Recently, weakly supervised methods still have a large performance gap compared to fully supervised ones, while the latter requires laborious timestamp annotations. In this study, we aim to reduce the annotation cost yet keep competitive performance for TSG task compared to fully supervised ones. To achieve this goal, we investigate a recently proposed glance-supervised temporal sentence grounding task, which requires only single frame annotation (referred to as glance annotation) for each query. Under this setup, we propose a Dynamic Gaussian prior based Grounding framework with Glance annotation (D3G), which consists of a Semantic Alignment Group Contrastive Learning module (SA-GCL) and a Dynamic Gaussian prior Adjustment module (DGA). Specifically, SA-GCL samples reliable positive moments from a 2D temporal map via jointly leveraging Gaussian prior and semantic consistency, which contributes to aligning the positive sentence-moment pairs in the joint embedding space. Moreover, to alleviate the annotation bias resulting from glance annotation and model complex queries consisting of multiple events, we propose the DGA module, which adjusts the distribution dynamically to approximate the ground truth of target moments. Extensive experiments on three challenging benchmarks verify the effectiveness of the proposed D3G. It outperforms the state-of-the-art weakly supervised methods by a large margin and narrows the performance gap compared to fully supervised methods. Code is available at https://github.com/solicucu/D3G.

Results

TaskDatasetMetricValueModel
Video UnderstandingCharades-STAR1@0.546D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR1@0.720.2D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR5@0.583.1D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR5@0.750.2D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR1@0.541.7D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR1@0.718.8D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR5@0.578.2D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Video UnderstandingCharades-STAR5@0.748D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR1@0.546D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR1@0.720.2D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR5@0.583.1D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR5@0.750.2D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR1@0.541.7D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR1@0.718.8D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR5@0.578.2D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
VideoCharades-STAR5@0.748D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR1@0.546D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR1@0.720.2D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR5@0.583.1D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR5@0.750.2D3G (Semi-weak, MViT-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR1@0.541.7D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR1@0.718.8D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR5@0.578.2D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)
Temporal Sentence GroundingCharades-STAR5@0.748D3G (Semi-weak, I3D-K400-Pretrain-feature, evaluated by AdaFocus)

Related Papers

SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16LLM-Driven Dual-Level Multi-Interest Modeling for Recommendation2025-07-15Latent Space Consistency for Sparse-View CT Reconstruction2025-07-15Self-supervised pretraining of vision transformers for animal behavioral analysis and neural encoding2025-07-13