TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Fast Abstractive Summarization with Reinforce-Selected Sen...

Fast Abstractive Summarization with Reinforce-Selected Sentence Rewriting

Yen-Chun Chen, Mohit Bansal

2018-05-28ACL 2018 7Abstractive Text SummarizationText SummarizationSentence ReWriting
PaperPDFCodeCodeCode(official)

Abstract

Inspired by how humans summarize long documents, we propose an accurate and fast summarization model that first selects salient sentences and then rewrites them abstractively (i.e., compresses and paraphrases) to generate a concise overall summary. We use a novel sentence-level policy gradient method to bridge the non-differentiable computation between these two neural networks in a hierarchical way, while maintaining language fluency. Empirically, we achieve the new state-of-the-art on all metrics (including human evaluation) on the CNN/Daily Mail dataset, as well as significantly higher abstractiveness scores. Moreover, by first operating at the sentence-level and then the word-level, we enable parallel decoding of our neural generative model that results in substantially faster (10-20x) inference speed as well as 4x faster training convergence than previous long-paragraph encoder-decoder models. We also demonstrate the generalization of our model on the test-only DUC-2002 dataset, where we achieve higher scores than a state-of-the-art model.

Results

TaskDatasetMetricValueModel
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-139.66rnn-ext + abs + RL + rerank
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-215.85rnn-ext + abs + RL + rerank
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-L37.34rnn-ext + abs + RL + rerank
Text SummarizationCNN / Daily MailROUGE-141.47rnn-ext + RL
Text SummarizationCNN / Daily MailROUGE-218.72rnn-ext + RL
Text SummarizationCNN / Daily MailROUGE-L37.76rnn-ext + RL
Text SummarizationCNN / Daily MailROUGE-140.88rnn-ext + abs + RL + rerank
Text SummarizationCNN / Daily MailROUGE-217.8rnn-ext + abs + RL + rerank
Text SummarizationCNN / Daily MailROUGE-L38.54rnn-ext + abs + RL + rerank
Abstractive Text SummarizationCNN / Daily MailROUGE-141.47rnn-ext + RL
Abstractive Text SummarizationCNN / Daily MailROUGE-218.72rnn-ext + RL
Abstractive Text SummarizationCNN / Daily MailROUGE-L37.76rnn-ext + RL
Abstractive Text SummarizationCNN / Daily MailROUGE-140.88rnn-ext + abs + RL + rerank
Abstractive Text SummarizationCNN / Daily MailROUGE-217.8rnn-ext + abs + RL + rerank
Abstractive Text SummarizationCNN / Daily MailROUGE-L38.54rnn-ext + abs + RL + rerank

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Improving large language models with concept-aware fine-tuning2025-06-09Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs2025-06-03ARC: Argument Representation and Coverage Analysis for Zero-Shot Long Document Summarization with Instruction Following LLMs2025-05-29MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection2025-05-29APE: A Data-Centric Benchmark for Efficient LLM Adaptation in Text Summarization2025-05-26FiLLM -- A Filipino-optimized Large Language Model based on Southeast Asia Large Language Model (SEALLM)2025-05-25