TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Coreferential Reasoning Learning for Language Representation

Coreferential Reasoning Learning for Language Representation

Deming Ye, Yankai Lin, Jiaju Du, Zheng-Hao Liu, Peng Li, Maosong Sun, Zhiyuan Liu

2020-04-15EMNLP 2020 11Relation Extraction
PaperPDFCode(official)Code(official)

Abstract

Language representation models such as BERT could effectively capture contextual semantic information from plain text, and have been proved to achieve promising results in lots of downstream NLP tasks with appropriate fine-tuning. However, most existing language representation models cannot explicitly handle coreference, which is essential to the coherent understanding of the whole discourse. To address this issue, we present CorefBERT, a novel language representation model that can capture the coreferential relations in context. The experimental results show that, compared with existing baseline models, CorefBERT can achieve significant improvements consistently on various downstream NLP tasks that require coreferential reasoning, while maintaining comparable performance to previous models on other common NLP tasks. The source code and experiment details of this paper can be obtained from https://github.com/thunlp/CorefBERT.

Results

TaskDatasetMetricValueModel
Relation ExtractionDocREDF160.25CorefRoBERTa-large
Relation ExtractionDocREDIgn F157.9CorefRoBERTa-large
Relation ExtractionDocREDF158.83CorefBERT-large
Relation ExtractionDocREDIgn F156.4CorefBERT-large
Relation ExtractionDocREDF156.96CorefBERT-base
Relation ExtractionDocREDIgn F154.54CorefBERT-base

Related Papers

DocIE@XLLM25: In-Context Learning for Information Extraction using Fully Synthetic Demonstrations2025-07-08Multiple Streams of Relation Extraction: Enriching and Recalling in Transformers2025-06-25Chaining Event Spans for Temporal Relation Grounding2025-06-17Summarization for Generative Relation Extraction in the Microbiome Domain2025-06-10Conservative Bias in Large Language Models: Measuring Relation Predictions2025-06-09Comparative Analysis of AI Agent Architectures for Entity Relationship Classification2025-06-03CREFT: Sequential Multi-Agent LLM for Character Relation Extraction2025-05-30Generating Diverse Training Samples for Relation Extraction with Large Language Models2025-05-29