TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Structured Prediction as Translation between Augmented Nat...

Structured Prediction as Translation between Augmented Natural Languages

Giovanni Paolini, Ben Athiwaratkun, Jason Krone, Jie Ma, Alessandro Achille, Rishita Anubhai, Cicero Nogueira dos santos, Bing Xiang, Stefano Soatto

2021-01-14ICLR 2021 1Nested Named Entity RecognitionStructured PredictionRelation Extractioncoreference-resolutionDialogue State TrackingCoreference Resolutionnamed-entity-recognitionNamed Entity RecognitionEvent ExtractionTranslationPredictionMulti-Task LearningSemantic Role LabelingJoint Entity and Relation ExtractionRelation ClassificationNamed Entity Recognition (NER)
PaperPDFCodeCode(official)

Abstract

We propose a new framework, Translation between Augmented Natural Languages (TANL), to solve many structured prediction language tasks including joint entity and relation extraction, nested named entity recognition, relation classification, semantic role labeling, event extraction, coreference resolution, and dialogue state tracking. Instead of tackling the problem by training task-specific discriminative classifiers, we frame it as a translation task between augmented natural languages, from which the task-relevant information can be easily extracted. Our approach can match or outperform task-specific models on all tasks, and in particular, achieves new state-of-the-art results on joint entity and relation extraction (CoNLL04, ADE, NYT, and ACE2005 datasets), relation classification (FewRel and TACRED), and semantic role labeling (CoNLL-2005 and CoNLL-2012). We accomplish this while using the same architecture and hyperparameters for all tasks and even when training a single model to solve all tasks at the same time (multi-task learning). Finally, we show that our framework can also significantly improve the performance in a low-resource regime, thanks to better use of label semantics.

Results

TaskDatasetMetricValueModel
Relation ExtractionTACREDF171.9TANL
Relation ExtractionCoNLL04RE+ Micro F172.6TANL
Relation ExtractionTACREDF161.9TANL (multi-task)
Relation ExtractionTACREDF171.9TANL
Relation ClassificationTACREDF161.9TANL (multi-task)
Relation ClassificationTACREDF171.9TANL

Related Papers

Multi-Strategy Improved Snake Optimizer Accelerated CNN-LSTM-Attention-Adaboost for Trajectory Prediction2025-07-21A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17SGCL: Unifying Self-Supervised and Supervised Learning for Graph Recommendation2025-07-17Function-to-Style Guidance of LLMs for Code Translation2025-07-15Generative Click-through Rate Prediction with Applications to Search Advertising2025-07-15Robust-Multi-Task Gradient Boosting2025-07-15Conformation-Aware Structure Prediction of Antigen-Recognizing Immune Proteins2025-07-11SAMO: A Lightweight Sharpness-Aware Approach for Multi-Task Optimization with Joint Global-Local Perturbation2025-07-10