TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Syntactically Look-Ahead Attention Network for Sentence Co...

Syntactically Look-Ahead Attention Network for Sentence Compression

Hidetaka Kamigaito, Manabu Okumura

2020-02-04Sentence CompressionInformativeness
PaperPDFCode(official)

Abstract

Sentence compression is the task of compressing a long sentence into a short one by deleting redundant words. In sequence-to-sequence (Seq2Seq) based models, the decoder unidirectionally decides to retain or delete words. Thus, it cannot usually explicitly capture the relationships between decoded words and unseen words that will be decoded in the future time steps. Therefore, to avoid generating ungrammatical sentences, the decoder sometimes drops important words in compressing sentences. To solve this problem, we propose a novel Seq2Seq model, syntactically look-ahead attention network (SLAHAN), that can generate informative summaries by explicitly tracking both dependency parent and child words during decoding and capturing important words that will be decoded in the future. The results of the automatic evaluation on the Google sentence compression dataset showed that SLAHAN achieved the best kept-token-based-F1, ROUGE-1, ROUGE-2 and ROUGE-L scores of 85.5, 79.3, 71.3 and 79.1, respectively. SLAHAN also improved the summarization performance on longer sentences. Furthermore, in the human evaluation, SLAHAN improved informativeness without losing readability.

Results

TaskDatasetMetricValueModel
Sentence EmbeddingsGoogle DatasetCR0.407SLAHAN (LSTM+syntactic-information)
Sentence EmbeddingsGoogle DatasetF10.855SLAHAN (LSTM+syntactic-information)
Text SummarizationGoogle DatasetCR0.407SLAHAN (LSTM+syntactic-information)
Text SummarizationGoogle DatasetF10.855SLAHAN (LSTM+syntactic-information)
Representation LearningGoogle DatasetCR0.407SLAHAN (LSTM+syntactic-information)
Representation LearningGoogle DatasetF10.855SLAHAN (LSTM+syntactic-information)
Sentence CompressionGoogle DatasetCR0.407SLAHAN (LSTM+syntactic-information)
Sentence CompressionGoogle DatasetF10.855SLAHAN (LSTM+syntactic-information)

Related Papers

Multi-Agent Retrieval-Augmented Framework for Evidence-Based Counterspeech Against Health Misinformation2025-07-09LumiCRS: Asymmetric Contrastive Prototype Learning for Long-Tail Conversational Movie Recommendation2025-07-07Dynamic Bandwidth Allocation for Hybrid Event-RGB Transmission2025-06-25Multi-Preference Lambda-weighted Listwise DPO for Dynamic Preference Alignment2025-06-24CuRe: Cultural Gaps in the Long Tail of Text-to-Image Systems2025-06-09Image Reconstruction as a Tool for Feature Analysis2025-06-09Investigating the Impact of Word Informativeness on Speech Emotion Recognition2025-06-02Assumption-free stability for ranking problems2025-06-02