TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/PEGASUS: Pre-training with Extracted Gap-sentences for Abs...

PEGASUS: Pre-training with Extracted Gap-sentences for Abstractive Summarization

Jingqing Zhang, Yao Zhao, Mohammad Saleh, Peter J. Liu

2019-12-18ICML 2020 1Abstractive Text SummarizationText Summarization
PaperPDFCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Recent work pre-training Transformers with self-supervised objectives on large text corpora has shown great success when fine-tuned on downstream NLP tasks including text summarization. However, pre-training objectives tailored for abstractive text summarization have not been explored. Furthermore there is a lack of systematic evaluation across diverse domains. In this work, we propose pre-training large Transformer-based encoder-decoder models on massive text corpora with a new self-supervised objective. In PEGASUS, important sentences are removed/masked from an input document and are generated together as one output sequence from the remaining sentences, similar to an extractive summary. We evaluated our best PEGASUS model on 12 downstream summarization tasks spanning news, science, stories, instructions, emails, patents, and legislative bills. Experiments demonstrate it achieves state-of-the-art performance on all 12 downstream datasets measured by ROUGE scores. Our model also shows surprising performance on low-resource summarization, surpassing previous state-of-the-art results on 6 datasets with only 1000 examples. Finally we validated our results using human evaluation and show that our model summaries achieve human performance on multiple datasets.

Results

TaskDatasetMetricValueModel
Text SummarizationGigaWordROUGE-139.12PEGASUS
Text SummarizationGigaWordROUGE-219.86PEGASUS
Text SummarizationGigaWordROUGE-L36.24PEGASUS
Text SummarizationArxiv HEP-TH citation graphROUGE-144.67PEGASUS
Text SummarizationPubmedROUGE-145.09PEGASUS
Text SummarizationX-SumROUGE-147.21PEGASUSLARGE
Text SummarizationX-SumROUGE-224.56PEGASUSLARGE
Text SummarizationCNN / Daily MailROUGE-144.17PEGASUS
Text SummarizationCNN / Daily MailROUGE-221.47PEGASUS
Text SummarizationCNN / Daily MailROUGE-L41.11PEGASUS
Text SummarizationAESLCROUGE-137.68PEGASUS
Text SummarizationAESLCROUGE-221.25PEGASUS
Text SummarizationAESLCROUGE-L36.51PEGASUS
Abstractive Text SummarizationCNN / Daily MailROUGE-144.17PEGASUS
Abstractive Text SummarizationCNN / Daily MailROUGE-221.47PEGASUS
Abstractive Text SummarizationCNN / Daily MailROUGE-L41.11PEGASUS
Abstractive Text SummarizationAESLCROUGE-137.68PEGASUS
Abstractive Text SummarizationAESLCROUGE-221.25PEGASUS
Abstractive Text SummarizationAESLCROUGE-L36.51PEGASUS

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Improving large language models with concept-aware fine-tuning2025-06-09Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs2025-06-03ARC: Argument Representation and Coverage Analysis for Zero-Shot Long Document Summarization with Instruction Following LLMs2025-05-29MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection2025-05-29APE: A Data-Centric Benchmark for Efficient LLM Adaptation in Text Summarization2025-05-26FiLLM -- A Filipino-optimized Large Language Model based on Southeast Asia Large Language Model (SEALLM)2025-05-25