TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Mixture Content Selection for Diverse Sequence Generation

Mixture Content Selection for Diverse Sequence Generation

Jaemin Cho, Minjoon Seo, Hannaneh Hajishirzi

2019-09-04IJCNLP 2019 11Abstractive Text SummarizationDocument SummarizationQuestion Generation
PaperPDFCode(official)

Abstract

Generating diverse sequences is important in many NLP applications such as question generation or summarization that exhibit semantically one-to-many relationships between source and the target sequences. We present a method to explicitly separate diversification from generation using a general plug-and-play module (called SELECTOR) that wraps around and guides an existing encoder-decoder model. The diversification stage uses a mixture of experts to sample different binary masks on the source sequence for diverse content selection. The generation stage uses a standard encoder-decoder model given each selected content from the source sequence. Due to the non-differentiable nature of discrete sampling and the lack of ground truth labels for binary mask, we leverage a proxy for ground truth mask and adopt stochastic hard-EM for training. In question generation (SQuAD) and abstractive summarization (CNN-DM), our method demonstrates significant improvements in accuracy, diversity and training efficiency, including state-of-the-art top-1 accuracy in both datasets, 6% gain in top-5 accuracy, and 3.7 times faster training over a state of the art model. Our code is publicly available at https://github.com/clovaai/FocusSeq2Seq.

Results

TaskDatasetMetricValueModel
Text SummarizationCNN / Daily MailROUGE-141.72Selector & Pointer-Generator
Text SummarizationCNN / Daily MailROUGE-218.74Selector & Pointer-Generator
Text SummarizationCNN / Daily MailROUGE-L38.79Selector & Pointer-Generator
Text SummarizationCNN / Daily MailROUGE-141.72Selector+Pointer Generator
Text SummarizationCNN / Daily MailROUGE-218.74Selector+Pointer Generator
Text SummarizationCNN / Daily MailROUGE-L38.79Selector+Pointer Generator
Abstractive Text SummarizationCNN / Daily MailROUGE-141.72Selector & Pointer-Generator
Abstractive Text SummarizationCNN / Daily MailROUGE-218.74Selector & Pointer-Generator
Abstractive Text SummarizationCNN / Daily MailROUGE-L38.79Selector & Pointer-Generator
Question GenerationSQuAD1.1BLEU-415.874Selector & NQG++
Document SummarizationCNN / Daily MailROUGE-141.72Selector+Pointer Generator
Document SummarizationCNN / Daily MailROUGE-218.74Selector+Pointer Generator
Document SummarizationCNN / Daily MailROUGE-L38.79Selector+Pointer Generator

Related Papers

Compressed and Smooth Latent Space for Text Diffusion Modeling2025-06-26GenerationPrograms: Fine-grained Attribution with Executable Programs2025-06-17ELLIS Alicante at CQs-Gen 2025: Winning the critical thinking questions shared task: LLM-based question generation and selection2025-06-17Arctic Long Sequence Training: Scalable And Efficient Training For Multi-Million Token Sequences2025-06-16Improving Fairness of Large Language Models in Multi-document Summarization2025-06-09Knowledge Compression via Question Generation: Enhancing Multihop Document Retrieval without Fine-tuning2025-06-09Multiple-Choice Question Generation Using Large Language Models: Methodology and Educator Insights2025-06-05Advancing Decoding Strategies: Enhancements in Locally Typical Sampling for LLMs2025-06-03