TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Neural Attention Model for Abstractive Sentence Summariz...

A Neural Attention Model for Abstractive Sentence Summarization

Alexander M. Rush, Sumit Chopra, Jason Weston

2015-09-02EMNLP 2015 9Extractive Text SummarizationText Summarization
PaperPDFCodeCodeCodeCode

Abstract

Summarization based on text extraction is inherently limited, but generation-style abstractive methods have proven challenging to build. In this work, we propose a fully data-driven approach to abstractive sentence summarization. Our method utilizes a local attention-based model that generates each word of the summary conditioned on the input sentence. While the model is structurally simple, it can easily be trained end-to-end and scales to a large amount of training data. The model shows significant performance gains on the DUC-2004 shared task compared with several strong baselines.

Results

TaskDatasetMetricValueModel
Text SummarizationDUC 2004 Task 1ROUGE-128.18Abs+
Text SummarizationDUC 2004 Task 1ROUGE-28.49Abs+
Text SummarizationDUC 2004 Task 1ROUGE-L23.81Abs+
Text SummarizationDUC 2004 Task 1ROUGE-L22.05ABS
Text SummarizationGigaWordROUGE-131Abs+
Text SummarizationGigaWordROUGE-130.88Abs
Text SummarizationDUC 2004 Task 1ROUGE-126.55Abs
Text SummarizationDUC 2004 Task 1ROUGE-27.06Abs
Text SummarizationDUC 2004 Task 1ROUGE-L22.05Abs
Extractive Text SummarizationDUC 2004 Task 1ROUGE-126.55Abs
Extractive Text SummarizationDUC 2004 Task 1ROUGE-27.06Abs
Extractive Text SummarizationDUC 2004 Task 1ROUGE-L22.05Abs

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Improving large language models with concept-aware fine-tuning2025-06-09MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection2025-05-29APE: A Data-Centric Benchmark for Efficient LLM Adaptation in Text Summarization2025-05-26FiLLM -- A Filipino-optimized Large Language Model based on Southeast Asia Large Language Model (SEALLM)2025-05-25Scaling Up Biomedical Vision-Language Models: Fine-Tuning, Instruction Tuning, and Multi-Modal Learning2025-05-23A Structured Literature Review on Traditional Approaches in Current Natural Language Processing2025-05-19