TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Extractive Summarization as Text Matching

Extractive Summarization as Text Matching

Ming Zhong, PengFei Liu, Yiran Chen, Danqing Wang, Xipeng Qiu, Xuanjing Huang

2020-04-19ACL 2020 6Text MatchingExtractive Text SummarizationText SummarizationDocument SummarizationExtractive SummarizationSemantic Text Matching
PaperPDFCode(official)Code

Abstract

This paper creates a paradigm shift with regard to the way we build neural extractive summarization systems. Instead of following the commonly used framework of extracting sentences individually and modeling the relationship between sentences, we formulate the extractive summarization task as a semantic text matching problem, in which a source document and candidate summaries will be (extracted from the original text) matched in a semantic space. Notably, this paradigm shift to semantic matching framework is well-grounded in our comprehensive analysis of the inherent gap between sentence-level and summary-level extractors based on the property of the dataset. Besides, even instantiating the framework with a simple form of a matching model, we have driven the state-of-the-art extractive result on CNN/DailyMail to a new level (44.41 in ROUGE-1). Experiments on the other five datasets also show the effectiveness of the matching framework. We believe the power of this matching-based summarization framework has not been fully exploited. To encourage more instantiations in the future, we have released our codes, processed dataset, as well as generated summaries in https://github.com/maszhongming/MatchSum.

Results

TaskDatasetMetricValueModel
Text SummarizationWikiHowROUGE-131.85MatchSum (BERT-base)
Text SummarizationWikiHowROUGE-28.98MatchSum (BERT-base)
Text SummarizationWikiHowROUGE-L29.58MatchSum (BERT-base)
Text SummarizationReddit TIFUROUGE-125.09MatchSum
Text SummarizationReddit TIFUROUGE-26.17MatchSum
Text SummarizationReddit TIFUROUGE-L20.13MatchSum
Text SummarizationBBC XSumROUGE-124.86MatchSum
Text SummarizationBBC XSumROUGE-24.66MatchSum
Text SummarizationBBC XSumROUGE-L18.41MatchSum
Text SummarizationPubmedROUGE-141.21MatchSum (BERT-base)
Text SummarizationPubmedROUGE-214.91MatchSum (BERT-base)
Text SummarizationPubmedROUGE-L36.75MatchSum (BERT-base)
Text SummarizationCNN / Daily MailROUGE-144.41MatchSum (RoBERTa-base)
Text SummarizationCNN / Daily MailROUGE-220.86MatchSum (RoBERTa-base)
Text SummarizationCNN / Daily MailROUGE-L40.55MatchSum (RoBERTa-base)
Text SummarizationCNN / Daily MailROUGE-144.22MatchSum (BERT-base)
Text SummarizationCNN / Daily MailROUGE-220.62MatchSum (BERT-base)
Text SummarizationCNN / Daily MailROUGE-L40.38MatchSum (BERT-base)
Text SummarizationCNN / Daily MailROUGE-144.41MatchSum
Text SummarizationCNN / Daily MailROUGE-220.86MatchSum
Text SummarizationCNN / Daily MailROUGE-L40.55MatchSum
Extractive Text SummarizationCNN / Daily MailROUGE-144.41MatchSum
Extractive Text SummarizationCNN / Daily MailROUGE-220.86MatchSum
Extractive Text SummarizationCNN / Daily MailROUGE-L40.55MatchSum
Document SummarizationCNN / Daily MailROUGE-144.41MatchSum (RoBERTa-base)
Document SummarizationCNN / Daily MailROUGE-220.86MatchSum (RoBERTa-base)
Document SummarizationCNN / Daily MailROUGE-L40.55MatchSum (RoBERTa-base)
Document SummarizationCNN / Daily MailROUGE-144.22MatchSum (BERT-base)
Document SummarizationCNN / Daily MailROUGE-220.62MatchSum (BERT-base)
Document SummarizationCNN / Daily MailROUGE-L40.38MatchSum (BERT-base)

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15GenerationPrograms: Fine-grained Attribution with Executable Programs2025-06-17Arctic Long Sequence Training: Scalable And Efficient Training For Multi-Million Token Sequences2025-06-16On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Efficient Medical Vision-Language Alignment Through Adapting Masked Vision Models2025-06-10Improving large language models with concept-aware fine-tuning2025-06-09Improving Fairness of Large Language Models in Multi-document Summarization2025-06-09MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection2025-05-29