TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Hierarchically Structured Reinforcement Learning for Topic...

Hierarchically Structured Reinforcement Learning for Topically Coherent Visual Story Generation

Qiuyuan Huang, Zhe Gan, Asli Celikyilmaz, Dapeng Wu, Jian-Feng Wang, Xiaodong He

2018-05-21Reinforcement LearningStory GenerationVisual Storytellingreinforcement-learning
PaperPDF

Abstract

We propose a hierarchically structured reinforcement learning approach to address the challenges of planning for generating coherent multi-sentence stories for the visual storytelling task. Within our framework, the task of generating a story given a sequence of images is divided across a two-level hierarchical decoder. The high-level decoder constructs a plan by generating a semantic concept (i.e., topic) for each image in sequence. The low-level decoder generates a sentence for each image using a semantic compositional network, which effectively grounds the sentence generation conditioned on the topic. The two decoders are jointly trained end-to-end using reinforcement learning. We evaluate our model on the visual storytelling (VIST) dataset. Empirical results from both automatic and human evaluations demonstrate that the proposed hierarchically structured reinforced training achieves significantly better performance compared to a strong flat deep reinforcement learning baseline.

Results

TaskDatasetMetricValueModel
Text GenerationVISTBLEU-412.32HSRL w/ Joint Training
Text GenerationVISTCIDEr10.71HSRL w/ Joint Training
Text GenerationVISTMETEOR35.23HSRL w/ Joint Training
Text GenerationVISTROUGE-L30.84HSRL w/ Joint Training
Text GenerationVISTSPICE12.97HSRL w/ Joint Training
Data-to-Text GenerationVISTBLEU-412.32HSRL w/ Joint Training
Data-to-Text GenerationVISTCIDEr10.71HSRL w/ Joint Training
Data-to-Text GenerationVISTMETEOR35.23HSRL w/ Joint Training
Data-to-Text GenerationVISTROUGE-L30.84HSRL w/ Joint Training
Data-to-Text GenerationVISTSPICE12.97HSRL w/ Joint Training
Visual StorytellingVISTBLEU-412.32HSRL w/ Joint Training
Visual StorytellingVISTCIDEr10.71HSRL w/ Joint Training
Visual StorytellingVISTMETEOR35.23HSRL w/ Joint Training
Visual StorytellingVISTROUGE-L30.84HSRL w/ Joint Training
Visual StorytellingVISTSPICE12.97HSRL w/ Joint Training
Story GenerationVISTBLEU-412.32HSRL w/ Joint Training
Story GenerationVISTCIDEr10.71HSRL w/ Joint Training
Story GenerationVISTMETEOR35.23HSRL w/ Joint Training
Story GenerationVISTROUGE-L30.84HSRL w/ Joint Training
Story GenerationVISTSPICE12.97HSRL w/ Joint Training

Related Papers

CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning2025-07-18VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Aligning Humans and Robots via Reinforcement Learning from Implicit Human Feedback2025-07-17VAR-MATH: Probing True Mathematical Reasoning in Large Language Models via Symbolic Multi-Instance Benchmarks2025-07-17QuestA: Expanding Reasoning Capacity in LLMs via Question Augmentation2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Autonomous Resource Management in Microservice Systems via Reinforcement Learning2025-07-17