TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Divide-and-Conquer Approach to the Summarization of Long...

A Divide-and-Conquer Approach to the Summarization of Long Documents

Alexios Gidiotis, Grigorios Tsoumakas

2020-04-13Sentence SimilarityText SummarizationDocument Summarization
PaperPDFCode(official)

Abstract

We present a novel divide-and-conquer method for the neural summarization of long documents. Our method exploits the discourse structure of the document and uses sentence similarity to split the problem into an ensemble of smaller summarization problems. In particular, we break a long document and its summary into multiple source-target pairs, which are used for training a model that learns to summarize each part of the document separately. These partial summaries are then combined in order to produce a final complete summary. With this approach we can decompose the problem of long document summarization into smaller and simpler problems, reducing computational complexity and creating more training examples, which at the same time contain less noise in the target summaries compared to the standard approach. We demonstrate that this approach paired with different summarization models, including sequence-to-sequence RNNs and Transformers, can lead to improved summarization performance. Our best models achieve results that are on par with the state-of-the-art in two two publicly available datasets of academic articles.

Results

TaskDatasetMetricValueModel
Text SummarizationArxiv HEP-TH citation graphROUGE-145.01DANCER PEGASUS
Text SummarizationArxiv HEP-TH citation graphROUGE-217.6DANCER PEGASUS
Text SummarizationArxiv HEP-TH citation graphROUGE-L40.56DANCER PEGASUS
Text SummarizationArxiv HEP-TH citation graphROUGE-142.7DANCER RUM
Text SummarizationArxiv HEP-TH citation graphROUGE-216.54DANCER RUM
Text SummarizationArxiv HEP-TH citation graphROUGE-L38.44DANCER RUM
Text SummarizationArxiv HEP-TH citation graphROUGE-141.87DANCER LSTM
Text SummarizationArxiv HEP-TH citation graphROUGE-215.92DANCER LSTM
Text SummarizationArxiv HEP-TH citation graphROUGE-L37.61DANCER LSTM
Text SummarizationPubmedROUGE-146.34DANCER PEGASUS
Text SummarizationPubmedROUGE-219.97DANCER PEGASUS
Text SummarizationPubmedROUGE-L42.42DANCER PEGASUS
Text SummarizationPubmedROUGE-144.09DANCER LSTM
Text SummarizationPubmedROUGE-217.69DANCER LSTM
Text SummarizationPubmedROUGE-L40.27DANCER LSTM
Text SummarizationPubmedROUGE-143.98DANCER RUM
Text SummarizationPubmedROUGE-217.65DANCER RUM
Text SummarizationPubmedROUGE-L40.25DANCER RUM

Related Papers

LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15GenerationPrograms: Fine-grained Attribution with Executable Programs2025-06-17Arctic Long Sequence Training: Scalable And Efficient Training For Multi-Million Token Sequences2025-06-16On-the-Fly Adaptive Distillation of Transformer to Dual-State Linear Attention2025-06-11Improving large language models with concept-aware fine-tuning2025-06-09Improving Fairness of Large Language Models in Multi-document Summarization2025-06-09EL4NER: Ensemble Learning for Named Entity Recognition via Multiple Small-Parameter Large Language Models2025-05-29MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection2025-05-29