TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Unified Model for Extractive and Abstractive Summarizati...

A Unified Model for Extractive and Abstractive Summarization using Inconsistency Loss

Wan-Ting Hsu, Chieh-Kai Lin, Ming-Ying Lee, Kerui Min, Jing Tang, Min Sun

2018-05-16ACL 2018 7Abstractive Text Summarization
PaperPDFCode

Abstract

We propose a unified model combining the strength of extractive and abstractive summarization. On the one hand, a simple extractive model can obtain sentence-level attention with high ROUGE scores but less readable. On the other hand, a more complicated abstractive model can obtain word-level dynamic attention to generate a more readable paragraph. In our model, sentence-level attention is used to modulate the word-level attention such that words in less attended sentences are less likely to be generated. Moreover, a novel inconsistency loss function is introduced to penalize the inconsistency between two levels of attentions. By end-to-end training our model with the inconsistency loss and original losses of extractive and abstractive models, we achieve state-of-the-art ROUGE scores while being the most informative and readable summarization on the CNN/Daily Mail dataset in a solid human evaluation.

Results

TaskDatasetMetricValueModel
Text SummarizationCNN / Daily MailROUGE-140.68end2end w/ inconsistency loss
Text SummarizationCNN / Daily MailROUGE-217.97end2end w/ inconsistency loss
Text SummarizationCNN / Daily MailROUGE-L37.13end2end w/ inconsistency loss
Abstractive Text SummarizationCNN / Daily MailROUGE-140.68end2end w/ inconsistency loss
Abstractive Text SummarizationCNN / Daily MailROUGE-217.97end2end w/ inconsistency loss
Abstractive Text SummarizationCNN / Daily MailROUGE-L37.13end2end w/ inconsistency loss

Related Papers

Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs2025-06-03ARC: Argument Representation and Coverage Analysis for Zero-Shot Long Document Summarization with Instruction Following LLMs2025-05-29Power-Law Decay Loss for Large Language Model Finetuning: Focusing on Information Sparsity to Enhance Generation Quality2025-05-22Enhancing Abstractive Summarization of Scientific Papers Using Structure Information2025-05-20Low-Resource Language Processing: An OCR-Driven Summarization and Translation Pipeline2025-05-16ProdRev: A DNN framework for empowering customers using generative pre-trained transformers2025-05-14A Split-then-Join Approach to Abstractive Summarization for Very Long Documents in a Low Resource Setting2025-05-11GASCADE: Grouped Summarization of Adverse Drug Event for Enhanced Cancer Pharmacovigilance2025-05-07