TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Summarization as Indirect Supervision for Relation Extract...

Summarization as Indirect Supervision for Relation Extraction

Keming Lu, I-Hung Hsu, Wenxuan Zhou, Mingyu Derek Ma, Muhao Chen

2022-05-19Relation Extraction
PaperPDFCode(official)

Abstract

Relation extraction (RE) models have been challenged by their reliance on training data with expensive annotations. Considering that summarization tasks aim at acquiring concise expressions of synoptical information from the longer context, these tasks naturally align with the objective of RE, i.e., extracting a kind of synoptical information that describes the relation of entity mentions. We present SuRE, which converts RE into a summarization formulation. SuRE leads to more precise and resource-efficient RE based on indirect supervision from summarization tasks. To achieve this goal, we develop sentence and relation conversion techniques that essentially bridge the formulation of summarization and RE tasks. We also incorporate constraint decoding techniques with Trie scoring to further enhance summarization-based RE with robust inference. Experiments on three RE datasets demonstrate the effectiveness of SuRE in both full-dataset and low-resource settings, showing that summarization is a promising source of indirect supervision to improve RE models.

Results

TaskDatasetMetricValueModel
Relation ExtractionTACREDF175.1SuRE (PEGASUS-large)
Relation ExtractionTACREDF1 (1% Few-Shot)52SuRE (PEGASUS-large)
Relation ExtractionTACREDF1 (10% Few-Shot)70.7SuRE (PEGASUS-large)
Relation ExtractionTACREDF1 (5% Few-Shot)64.9SuRE (PEGASUS-large)
Relation ExtractionTACREDF1 (Zero-Shot)20.6SuRE (PEGASUS-large)

Related Papers

DocIE@XLLM25: In-Context Learning for Information Extraction using Fully Synthetic Demonstrations2025-07-08Multiple Streams of Relation Extraction: Enriching and Recalling in Transformers2025-06-25Chaining Event Spans for Temporal Relation Grounding2025-06-17Summarization for Generative Relation Extraction in the Microbiome Domain2025-06-10Conservative Bias in Large Language Models: Measuring Relation Predictions2025-06-09Comparative Analysis of AI Agent Architectures for Entity Relationship Classification2025-06-03CREFT: Sequential Multi-Agent LLM for Character Relation Extraction2025-05-30Generating Diverse Training Samples for Relation Extraction with Large Language Models2025-05-29