TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MaskGAN: Better Text Generation via Filling in the______

MaskGAN: Better Text Generation via Filling in the______

William Fedus, Ian Goodfellow, Andrew M. Dai

2018-01-23Text GenerationMultivariate Time Series Imputation
PaperPDF

Abstract

Neural text generation models are often autoregressive language models or seq2seq models. These models generate text by sampling words sequentially, with each word conditioned on the previous word, and are state-of-the-art for several machine translation and summarization benchmarks. These benchmarks are often defined by validation perplexity even though this is not a direct measure of the quality of the generated text. Additionally, these models are typically trained via maxi- mum likelihood and teacher forcing. These methods are well-suited to optimizing perplexity but can result in poor sample quality since generating text requires conditioning on sequences of words that may have never been observed at training time. We propose to improve sample quality using Generative Adversarial Networks (GANs), which explicitly train the generator to produce high quality samples and have shown a lot of success in image generation. GANs were originally designed to output differentiable values, so discrete language generation is challenging for them. We claim that validation perplexity alone is not indicative of the quality of text generated by a model. We introduce an actor-critic conditional GAN that fills in missing text conditioned on the surrounding context. We show qualitatively and quantitatively, evidence that this produces more realistic conditional and unconditional text samples compared to a maximum likelihood trained model.

Results

TaskDatasetMetricValueModel
ImputationBasketball Players MovementOOB Rate (10^−3) 4.592MaskGAN
ImputationBasketball Players MovementPath Difference0.68MaskGAN
ImputationBasketball Players MovementPath Length0.793MaskGAN
ImputationBasketball Players MovementPlayer Distance 0.427MaskGAN
ImputationBasketball Players MovementStep Change (10^−3)9.622MaskGAN
Feature EngineeringBasketball Players MovementOOB Rate (10^−3) 4.592MaskGAN
Feature EngineeringBasketball Players MovementPath Difference0.68MaskGAN
Feature EngineeringBasketball Players MovementPath Length0.793MaskGAN
Feature EngineeringBasketball Players MovementPlayer Distance 0.427MaskGAN
Feature EngineeringBasketball Players MovementStep Change (10^−3)9.622MaskGAN

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs2025-07-15Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Exploiting Leaderboards for Large-Scale Distribution of Malicious Models2025-07-11CLI-RAG: A Retrieval-Augmented Framework for Clinically Structured and Context Aware Text Generation with LLMs2025-07-09FIFA: Unified Faithfulness Evaluation Framework for Text-to-Video and Video-to-Text Generation2025-07-09