TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SummaRuNNer: A Recurrent Neural Network based Sequence Mod...

SummaRuNNer: A Recurrent Neural Network based Sequence Model for Extractive Summarization of Documents

Ramesh Nallapati, FeiFei Zhai, Bo-Wen Zhou

2016-11-14Text SummarizationDocument SummarizationExtractive Summarization
PaperPDFCodeCodeCodeCodeCodeCodeCode

Abstract

We present SummaRuNNer, a Recurrent Neural Network (RNN) based sequence model for extractive summarization of documents and show that it achieves performance better than or comparable to state-of-the-art. Our model has the additional advantage of being very interpretable, since it allows visualization of its predictions broken up by abstract features such as information content, salience and novelty. Another novel contribution of our work is abstractive training of our extractive model that can train on human generated reference summaries alone, eliminating the need for sentence-level extractive labels.

Results

TaskDatasetMetricValueModel
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-139.6SummaRuNNer
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-216.2SummaRuNNer
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-L35.3SummaRuNNer
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-139.2Lead-3 baseline
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-215.7Lead-3 baseline
Text SummarizationCNN / Daily Mail (Anonymized)ROUGE-L35.5Lead-3 baseline

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15GenerationPrograms: Fine-grained Attribution with Executable Programs2025-06-17Arctic Long Sequence Training: Scalable And Efficient Training For Multi-Million Token Sequences2025-06-16On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Improving large language models with concept-aware fine-tuning2025-06-09Improving Fairness of Large Language Models in Multi-document Summarization2025-06-09MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection2025-05-29ARC: Argument Representation and Coverage Analysis for Zero-Shot Long Document Summarization with Instruction Following LLMs2025-05-29