TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Text-to-Text Pre-Training for Data-to-Text Tasks

Text-to-Text Pre-Training for Data-to-Text Tasks

Mihir Kale, Abhinav Rastogi

2020-05-21INLG (ACL) 2020 12Data-to-Text GenerationText GenerationTransfer LearningLanguage Modelling
PaperPDFCode(official)Code

Abstract

We study the pre-train + fine-tune strategy for data-to-text tasks. Our experiments indicate that text-to-text pre-training in the form of T5, enables simple, end-to-end transformer based models to outperform pipelined neural architectures tailored for data-to-text generation, as well as alternative language model based pre-training techniques such as BERT and GPT-2. Importantly, T5 pre-training leads to better generalization, as evidenced by large improvements on out-of-domain test sets. We hope our work serves as a useful baseline for future research, as transfer learning becomes ever more prevalent for data-to-text tasks.

Results

TaskDatasetMetricValueModel
Text GenerationWebNLGBLEU64.7T5-Base
Text GenerationMULTIWOZ 2.1BLEU35.1T5-Base
Text GenerationWebNLG FullBLEU57.1T5-Large
Text GenerationToTToBLEU49.5T5-3B
Text GenerationToTToPARENT58.4T5-3B
Data-to-Text GenerationWebNLGBLEU64.7T5-Base
Data-to-Text GenerationMULTIWOZ 2.1BLEU35.1T5-Base
Data-to-Text GenerationWebNLG FullBLEU57.1T5-Large
Data-to-Text GenerationToTToBLEU49.5T5-3B
Data-to-Text GenerationToTToPARENT58.4T5-3B

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Making Language Model a Hierarchical Classifier and Generator2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16