TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Long Text Generation via Adversarial Training with Leaked ...

Long Text Generation via Adversarial Training with Leaked Information

Jiaxian Guo, Sidi Lu, Han Cai, Wei-Nan Zhang, Yong Yu, Jun Wang

2017-09-24Text GenerationReinforcement Learning
PaperPDFCodeCodeCodeCodeCode(official)Code

Abstract

Automatically generating coherent and semantically meaningful text has many applications in machine translation, dialogue systems, image captioning, etc. Recently, by combining with policy gradient, Generative Adversarial Nets (GAN) that use a discriminative model to guide the training of the generative model as a reinforcement learning policy has shown promising results in text generation. However, the scalar guiding signal is only available after the entire text has been generated and lacks intermediate information about text structure during the generative process. As such, it limits its success when the length of the generated text samples is long (more than 20 words). In this paper, we propose a new framework, called LeakGAN, to address the problem for long text generation. We allow the discriminative net to leak its own high-level extracted features to the generative net to further help the guidance. The generator incorporates such informative signals into all generation steps through an additional Manager module, which takes the extracted features of current generated words and outputs a latent vector to guide the Worker module for next-word generation. Our extensive experiments on synthetic data and various real-world tasks with Turing test demonstrate that LeakGAN is highly effective in long text generation and also improves the performance in short text generation scenarios. More importantly, without any supervision, LeakGAN would be able to implicitly learn sentence structures only through the interaction between Manager and Worker.

Results

TaskDatasetMetricValueModel
Text GenerationCOCO CaptionsBLEU-20.95LeakGAN
Text GenerationCOCO CaptionsBLEU-30.88LeakGAN
Text GenerationCOCO CaptionsBLEU-40.778LeakGAN
Text GenerationCOCO CaptionsBLEU-50.686LeakGAN
Text GenerationCOCO CaptionsBLEU-20.91partGAN
Text GenerationCOCO CaptionsBLEU-30.713partGAN
Text GenerationCOCO CaptionsBLEU-50.59partGAN
Text GenerationEMNLP2017 WMTBLEU-20.956LeakGAN
Text GenerationEMNLP2017 WMTBLEU-30.819LeakGAN
Text GenerationEMNLP2017 WMTBLEU-40.627LeakGAN
Text GenerationEMNLP2017 WMTBLEU-50.498LeakGAN
Text GenerationChinese PoemsBLEU-20.456LeakGAN

Related Papers

CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning2025-07-18Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Aligning Humans and Robots via Reinforcement Learning from Implicit Human Feedback2025-07-17VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17