TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Entity Commonsense Representation for Neural Abstractive S...

Entity Commonsense Representation for Neural Abstractive Summarization

Reinald Kim Amplayo, Seonjae Lim, Seung-won Hwang

2018-06-14NAACL 2018 6Abstractive Text SummarizationEntity Linking
PaperPDFCode(official)

Abstract

A major proportion of a text summary includes important entities found in the original text. These entities build up the topic of the summary. Moreover, they hold commonsense information once they are linked to a knowledge base. Based on these observations, this paper investigates the usage of linked entities to guide the decoder of a neural text summarizer to generate concise and better summaries. To this end, we leverage on an off-the-shelf entity linking system (ELS) to extract linked entities and propose Entity2Topic (E2T), a module easily attachable to a sequence-to-sequence model that transforms a list of entities into a vector representation of the topic of the summary. Current available ELS's are still not sufficiently effective, possibly introducing unresolved ambiguities and irrelevant entities. We resolve the imperfections of the ELS by (a) encoding entities with selective disambiguation, and (b) pooling entity vectors using firm attention. By applying E2T to a simple sequence-to-sequence model with attention mechanism as base model, we see significant improvements of the performance in the Gigaword (sentence to title) and CNN (long document to multi-sentence highlights) summarization datasets by at least 2 ROUGE points.

Results

TaskDatasetMetricValueModel
Text SummarizationGigaWordROUGE-137.04Seq2seq + E2T_cnn
Text SummarizationGigaWordROUGE-216.66Seq2seq + E2T_cnn
Text SummarizationGigaWordROUGE-L34.93Seq2seq + E2T_cnn

Related Papers

Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs2025-06-03LEMONADE: A Large Multilingual Expert-Annotated Abstractive Event Dataset for the Real World2025-06-01ARC: Argument Representation and Coverage Analysis for Zero-Shot Long Document Summarization with Instruction Following LLMs2025-05-29Distilling Closed-Source LLM's Knowledge for Locally Stable and Economic Biomedical Entity Linking2025-05-26Power-Law Decay Loss for Large Language Model Finetuning: Focusing on Information Sparsity to Enhance Generation Quality2025-05-22Enhancing Abstractive Summarization of Scientific Papers Using Structure Information2025-05-20Low-Resource Language Processing: An OCR-Driven Summarization and Translation Pipeline2025-05-16ProdRev: A DNN framework for empowering customers using generative pre-trained transformers2025-05-14