TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Eider: Empowering Document-level Relation Extraction with ...

Eider: Empowering Document-level Relation Extraction with Efficient Evidence Extraction and Inference-stage Fusion

Yiqing Xie, Jiaming Shen, Sha Li, Yuning Mao, Jiawei Han

2021-06-16Findings (ACL) 2022 5Relation ExtractionDocument-level Relation Extraction
PaperPDFCode(official)

Abstract

Document-level relation extraction (DocRE) aims to extract semantic relations among entity pairs in a document. Typical DocRE methods blindly take the full document as input, while a subset of the sentences in the document, noted as the evidence, are often sufficient for humans to predict the relation of an entity pair. In this paper, we propose an evidence-enhanced framework, Eider, that empowers DocRE by efficiently extracting evidence and effectively fusing the extracted evidence in inference. We first jointly train an RE model with a lightweight evidence extraction model, which is efficient in both memory and runtime. Empirically, even training the evidence model on silver labels constructed by our heuristic rules can lead to better RE performance. We further design a simple yet effective inference process that makes RE predictions on both extracted evidence and the full document, then fuses the predictions through a blending layer. This allows Eider to focus on important sentences while still having access to the complete information in the document. Extensive experiments show that Eider outperforms state-of-the-art methods on three benchmark datasets (e.g., by 1.37/1.26 Ign F1/F1 on DocRED).

Results

TaskDatasetMetricValueModel
Relation ExtractionDocREDF164.79Eider-RoBERTa-large
Relation ExtractionDocREDIgn F162.85Eider-RoBERTa-large
Relation ExtractionDocREDF162.47Eider-BERT-base
Relation ExtractionDocREDIgn F160.42Eider-BERT-base

Related Papers

DocIE@XLLM25: In-Context Learning for Information Extraction using Fully Synthetic Demonstrations2025-07-08Multiple Streams of Relation Extraction: Enriching and Recalling in Transformers2025-06-25Chaining Event Spans for Temporal Relation Grounding2025-06-17Summarization for Generative Relation Extraction in the Microbiome Domain2025-06-10Conservative Bias in Large Language Models: Measuring Relation Predictions2025-06-09Comparative Analysis of AI Agent Architectures for Entity Relationship Classification2025-06-03CREFT: Sequential Multi-Agent LLM for Character Relation Extraction2025-05-30Generating Diverse Training Samples for Relation Extraction with Large Language Models2025-05-29