TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Get To The Point: Summarization with Pointer-Generator Net...

Get To The Point: Summarization with Pointer-Generator Networks

Abigail See, Peter J. Liu, Christopher D. Manning

2017-04-14ACL 2017 7Abstractive Text SummarizationExtractive Text SummarizationText SummarizationDocument Summarization
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Neural sequence-to-sequence models have provided a viable new approach for abstractive text summarization (meaning they are not restricted to simply selecting and rearranging passages from the original text). However, these models have two shortcomings: they are liable to reproduce factual details inaccurately, and they tend to repeat themselves. In this work we propose a novel architecture that augments the standard sequence-to-sequence attentional model in two orthogonal ways. First, we use a hybrid pointer-generator network that can copy words from the source text via pointing, which aids accurate reproduction of information, while retaining the ability to produce novel words through the generator. Second, we use coverage to keep track of what has been summarized, which discourages repetition. We apply our model to the CNN / Daily Mail summarization task, outperforming the current abstractive state-of-the-art by at least 2 ROUGE points.

Results

TaskDatasetMetricValueModel
Text SummarizationArxiv HEP-TH citation graphROUGE-132.06Pntr-Gen-Seq2Seq
Text SummarizationPubmedROUGE-135.86Pntr-Gen-Seq2Seq
Text SummarizationCNN / Daily MailROUGE-139.53PTGEN + Coverage
Text SummarizationCNN / Daily MailROUGE-217.28PTGEN + Coverage
Text SummarizationCNN / Daily MailROUGE-L36.38PTGEN + Coverage
Text SummarizationCNN / Daily MailROUGE-139.53PTGEN + Coverage
Text SummarizationCNN / Daily MailROUGE-217.28PTGEN + Coverage
Text SummarizationCNN / Daily MailROUGE-L36.38PTGEN + Coverage
Text SummarizationCNN / Daily MailROUGE-139.53Pointer-Generator + Coverage
Text SummarizationCNN / Daily MailROUGE-217.28Pointer-Generator + Coverage
Text SummarizationCNN / Daily MailROUGE-140.34Lead-3
Text SummarizationCNN / Daily MailROUGE-217.7Lead-3
Text SummarizationCNN / Daily MailROUGE-L36.57Lead-3
Text SummarizationCNN / Daily MailROUGE-140.34Lead-3 baseline
Text SummarizationCNN / Daily MailROUGE-217.7Lead-3 baseline
Text SummarizationCNN / Daily MailROUGE-L36.57Lead-3 baseline
Extractive Text SummarizationCNN / Daily MailROUGE-140.34Lead-3 baseline
Extractive Text SummarizationCNN / Daily MailROUGE-217.7Lead-3 baseline
Extractive Text SummarizationCNN / Daily MailROUGE-L36.57Lead-3 baseline
Abstractive Text SummarizationCNN / Daily MailROUGE-139.53PTGEN + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-217.28PTGEN + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-L36.38PTGEN + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-139.53PTGEN + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-217.28PTGEN + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-L36.38PTGEN + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-139.53Pointer-Generator + Coverage
Abstractive Text SummarizationCNN / Daily MailROUGE-217.28Pointer-Generator + Coverage
Document SummarizationCNN / Daily MailROUGE-140.34Lead-3
Document SummarizationCNN / Daily MailROUGE-217.7Lead-3
Document SummarizationCNN / Daily MailROUGE-L36.57Lead-3

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15GenerationPrograms: Fine-grained Attribution with Executable Programs2025-06-17Arctic Long Sequence Training: Scalable And Efficient Training For Multi-Million Token Sequences2025-06-16On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Improving large language models with concept-aware fine-tuning2025-06-09Improving Fairness of Large Language Models in Multi-document Summarization2025-06-09Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs2025-06-03ARC: Argument Representation and Coverage Analysis for Zero-Shot Long Document Summarization with Instruction Following LLMs2025-05-29