TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Zero-Shot Information Extraction as a Unified Text-to-Trip...

Zero-Shot Information Extraction as a Unified Text-to-Triple Translation

Chenguang Wang, Xiao Liu, Zui Chen, Haoyun Hong, Jie Tang, Dawn Song

2021-09-23EMNLP 2021 11TranslationOpen Information ExtractionRelation ClassificationFactual probeLanguage Modelling
PaperPDFCode(official)

Abstract

We cast a suite of information extraction tasks into a text-to-triple translation framework. Instead of solving each task relying on task-specific datasets and models, we formalize the task as a translation between task-specific input text and output triples. By taking the task-specific input, we enable a task-agnostic translation by leveraging the latent knowledge that a pre-trained language model has about the task. We further demonstrate that a simple pre-training task of predicting which relational information corresponds to which input text is an effective way to produce task-specific outputs. This enables the zero-shot transfer of our framework to downstream tasks. We study the zero-shot performance of this framework on open information extraction (OIE2016, NYT, WEB, PENN), relation classification (FewRel and TACRED), and factual probe (Google-RE and T-REx). The model transfers non-trivially to most tasks and is often competitive with a fully supervised method without the need for any task-specific training. For instance, we significantly outperform the F1 score of the supervised open information extraction without needing to use its training set.

Results

TaskDatasetMetricValueModel
Relation ExtractionTACREDF149.2DeepEx (zero-shot top-1)
Relation ExtractionTACREDF176.4DeepEx (zero-shot top-10)
Relation ExtractionFewRelF148.8DeepEx (zero-shot top-1)
Relation ExtractionFewRelF192.9DeepEx (zero-shot top-10)
Relation ClassificationTACREDF149.2DeepEx (zero-shot top-1)
Relation ClassificationTACREDF176.4DeepEx (zero-shot top-10)
Relation ClassificationFewRelF148.8DeepEx (zero-shot top-1)
Relation ClassificationFewRelF192.9DeepEx (zero-shot top-10)
Open Information ExtractionWebAUC82.4DeepEx (zero-shot)
Open Information ExtractionWebF191.2DeepEx (zero-shot)
Open Information ExtractionPenn TreebankAUC81.5DeepEx (zero-shot)
Open Information ExtractionPenn TreebankF188.5DeepEx (zero-shot)
Open Information ExtractionNYTAUC72.5DeepEx (zero-shot)
Open Information ExtractionNYTF185.5DeepEx (zero-shot)
Open Information ExtractionOIE2016AUC58.6DeepEx (zero-shot)
Open Information ExtractionOIE2016F172.6DeepEx (zero-shot)

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Assay2Mol: large language model-based drug design using BioAssay context2025-07-16Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16