TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Salient Information Prompting to Steer Content in Prompt-b...

Salient Information Prompting to Steer Content in Prompt-based Abstractive Summarization

Lei Xu, Mohammed Asad Karim, Saket Dingliwal, Aparna Elangovan

2024-10-03Abstractive Text SummarizationText SummarizationHallucination
PaperPDFCode(official)

Abstract

Large language models (LLMs) can generate fluent summaries across domains using prompting techniques, reducing the need to train models for summarization applications. However, crafting effective prompts that guide LLMs to generate summaries with the appropriate level of detail and writing style remains a challenge. In this paper, we explore the use of salient information extracted from the source document to enhance summarization prompts. We show that adding keyphrases in prompts can improve ROUGE F1 and recall, making the generated summaries more similar to the reference and more complete. The number of keyphrases can control the precision-recall trade-off. Furthermore, our analysis reveals that incorporating phrase-level salient information is superior to word- or sentence-level. However, the impact on hallucination is not universally positive across LLMs. To conduct this analysis, we introduce Keyphrase Signal Extractor (SigExt), a lightweight model that can be finetuned to extract salient keyphrases. By using SigExt, we achieve consistent ROUGE improvements across datasets and open-weight and proprietary LLMs without any LLM customization. Our findings provide insights into leveraging salient information in building prompt-based summarization systems. We release our code at \url{https://github.com/amazon-science/SigExt}

Results

TaskDatasetMetricValueModel
Text SummarizationarXiv Summarization DatasetROUGE-145.2Claude Instant + SigExt
Text SummarizationarXiv Summarization DatasetROUGE-L23.5Claude Instant + SigExt
Text SummarizationSAMSumROUGE-144.1Mistral 7B + SigExt
Text SummarizationSAMSumROUGE-L33.9Mistral 7B + SigExt
Text SummarizationMeetingBankROUGE-L31.9Claude Instant + SigExt
Text SummarizationMeetingBankRouge-142.3Claude Instant + SigExt
Text SummarizationCNN/Daily MailROUGE-142Claude Instant + SigExt
Text SummarizationCNN/Daily MailROUGE-L26.6Claude Instant + SigExt
Abstractive Text SummarizationCNN/Daily MailROUGE-142Claude Instant + SigExt
Abstractive Text SummarizationCNN/Daily MailROUGE-L26.6Claude Instant + SigExt

Related Papers

Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way2025-07-11UQLM: A Python Package for Uncertainty Quantification in Large Language Models2025-07-08DeepRetro: Retrosynthetic Pathway Discovery using Iterative LLM Reasoning2025-07-07ReLoop: "Seeing Twice and Thinking Backwards" via Closed-loop Training to Mitigate Hallucinations in Multimodal understanding2025-07-07The Future is Agentic: Definitions, Perspectives, and Open Challenges of Multi-Agent Recommender Systems2025-07-02GAF-Guard: An Agentic Framework for Risk Management and Governance in Large Language Models2025-07-01