TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SPOT: Knowledge-Enhanced Language Representations for Info...

SPOT: Knowledge-Enhanced Language Representations for Information Extraction

Jiacheng Li, Yannis Katsis, Tyler Baldwin, Ho-Cheol Kim, Andrew Bartko, Julian McAuley, Chun-Nan Hsu

2022-08-20Relation Extraction
PaperPDF

Abstract

Knowledge-enhanced pre-trained models for language representation have been shown to be more effective in knowledge base construction tasks (i.e.,~relation extraction) than language models such as BERT. These knowledge-enhanced language models incorporate knowledge into pre-training to generate representations of entities or relationships. However, existing methods typically represent each entity with a separate embedding. As a result, these methods struggle to represent out-of-vocabulary entities and a large amount of parameters, on top of their underlying token models (i.e.,~the transformer), must be used and the number of entities that can be handled is limited in practice due to memory constraints. Moreover, existing models still struggle to represent entities and relationships simultaneously. To address these problems, we propose a new pre-trained model that learns representations of both entities and relationships from token spans and span pairs in the text respectively. By encoding spans efficiently with span modules, our model can represent both entities and their relationships but requires fewer parameters than existing models. We pre-trained our model with the knowledge graph extracted from Wikipedia and test it on a broad range of supervised and unsupervised information extraction tasks. Results show that our model learns better representations for both entities and relationships than baselines, while in supervised settings, fine-tuning our model outperforms RoBERTa consistently and achieves competitive results on information extraction tasks.

Results

TaskDatasetMetricValueModel
Relation ExtractionSemEval-2010 Task-8F190.6SPOT
Relation ExtractionSemEval-2010 Task-8F190.3LUKE
Relation ExtractionSemEval-2010 Task-8F189.4BERT
Relation ExtractionSemEval-2010 Task-8F189.2CorefBERT
Relation ExtractionSemEval-2010 Task-8F189.1KnowBERT
Relation ExtractionSemEval-2010 Task-8F188.8SpanBERT
Relation ExtractionSemEval-2010 Task-8F188.7RoBERTa

Related Papers

DocIE@XLLM25: In-Context Learning for Information Extraction using Fully Synthetic Demonstrations2025-07-08Multiple Streams of Relation Extraction: Enriching and Recalling in Transformers2025-06-25Chaining Event Spans for Temporal Relation Grounding2025-06-17Summarization for Generative Relation Extraction in the Microbiome Domain2025-06-10Conservative Bias in Large Language Models: Measuring Relation Predictions2025-06-09Comparative Analysis of AI Agent Architectures for Entity Relationship Classification2025-06-03CREFT: Sequential Multi-Agent LLM for Character Relation Extraction2025-05-30Generating Diverse Training Samples for Relation Extraction with Large Language Models2025-05-29