TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/ERNIE-GEN: An Enhanced Multi-Flow Pre-training and Fine-tu...

ERNIE-GEN: An Enhanced Multi-Flow Pre-training and Fine-tuning Framework for Natural Language Generation

Dongling Xiao, Han Zhang, Yukun Li, Yu Sun, Hao Tian, Hua Wu, Haifeng Wang

2020-01-26Text GenerationAbstractive Text SummarizationText SummarizationGenerative Question AnsweringDialogue GenerationQuestion Generation
PaperPDFCodeCodeCodeCodeCode

Abstract

Current pre-training works in natural language generation pay little attention to the problem of exposure bias on downstream tasks. To address this issue, we propose an enhanced multi-flow sequence to sequence pre-training and fine-tuning framework named ERNIE-GEN, which bridges the discrepancy between training and inference with an infilling generation mechanism and a noise-aware generation method. To make generation closer to human writing patterns, this framework introduces a span-by-span generation flow that trains the model to predict semantically-complete spans consecutively rather than predicting word by word. Unlike existing pre-training methods, ERNIE-GEN incorporates multi-granularity target sampling to construct pre-training data, which enhances the correlation between encoder and decoder. Experimental results demonstrate that ERNIE-GEN achieves state-of-the-art results with a much smaller amount of pre-training data and parameters on a range of language generation tasks, including abstractive summarization (Gigaword and CNN/DailyMail), question generation (SQuAD), dialogue generation (Persona-Chat) and generative question answering (CoQA).

Results

TaskDatasetMetricValueModel
Question AnsweringCoQAF1-Score84.5ERNIE-GEN
Text SummarizationGigaWordROUGE-139.46ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationGigaWordROUGE-220.34ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationGigaWordROUGE-L36.74ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationGigaWordROUGE-139.25ERNIE-GENLARGE
Text SummarizationGigaWordROUGE-220.25ERNIE-GENLARGE
Text SummarizationGigaWordROUGE-L36.53ERNIE-GENLARGE
Text SummarizationGigaWordROUGE-138.83ERNIE-GENBASE
Text SummarizationGigaWordROUGE-220.04ERNIE-GENBASE
Text SummarizationGigaWordROUGE-L36.2ERNIE-GENBASE
Text SummarizationGigaWord-10kROUGE-135.51ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationGigaWord-10kROUGE-216.79ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationGigaWord-10kROUGE-L33.23ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationGigaWord-10kROUGE-135.05ERNIE-GENLARGE
Text SummarizationGigaWord-10kROUGE-216.1ERNIE-GENLARGE
Text SummarizationGigaWord-10kROUGE-L32.5ERNIE-GENLARGE
Text SummarizationGigaWord-10kROUGE-133.75ERNIE-GENBASE
Text SummarizationGigaWord-10kROUGE-215.23ERNIE-GENBASE
Text SummarizationGigaWord-10kROUGE-L31.35ERNIE-GENBASE
Text SummarizationCNN / Daily MailROUGE-144.31ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationCNN / Daily MailROUGE-221.35ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationCNN / Daily MailROUGE-L41.6ERNIE-GENLARGE (large-scale text corpora)
Text SummarizationCNN / Daily MailROUGE-144.02ERNIE-GENLARGE
Text SummarizationCNN / Daily MailROUGE-221.17ERNIE-GENLARGE
Text SummarizationCNN / Daily MailROUGE-L41.26ERNIE-GENLARGE
Text SummarizationCNN / Daily MailROUGE-142.3ERNIE-GENBASE
Text SummarizationCNN / Daily MailROUGE-219.92ERNIE-GENBASE
Text SummarizationCNN / Daily MailROUGE-L39.68ERNIE-GENBASE
Abstractive Text SummarizationCNN / Daily MailROUGE-144.31ERNIE-GENLARGE (large-scale text corpora)
Abstractive Text SummarizationCNN / Daily MailROUGE-221.35ERNIE-GENLARGE (large-scale text corpora)
Abstractive Text SummarizationCNN / Daily MailROUGE-L41.6ERNIE-GENLARGE (large-scale text corpora)
Abstractive Text SummarizationCNN / Daily MailROUGE-144.02ERNIE-GENLARGE
Abstractive Text SummarizationCNN / Daily MailROUGE-221.17ERNIE-GENLARGE
Abstractive Text SummarizationCNN / Daily MailROUGE-L41.26ERNIE-GENLARGE
Abstractive Text SummarizationCNN / Daily MailROUGE-142.3ERNIE-GENBASE
Abstractive Text SummarizationCNN / Daily MailROUGE-219.92ERNIE-GENBASE
Abstractive Text SummarizationCNN / Daily MailROUGE-L39.68ERNIE-GENBASE
Question GenerationSQuAD1.1BLEU-425.41ERNIE-GENLARGE (beam size=5)

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17Emotional Support with LLM-based Empathetic Dialogue Generation2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs2025-07-15Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15ZipVoice-Dialog: Non-Autoregressive Spoken Dialogue Generation with Flow Matching2025-07-12