TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Stochastic Answer Networks for Machine Reading Comprehension

Stochastic Answer Networks for Machine Reading Comprehension

Xiaodong Liu, Yelong Shen, Kevin Duh, Jianfeng Gao

2017-12-10ACL 2018 7Reading ComprehensionQuestion AnsweringReinforcement LearningMachine Reading Comprehensionreinforcement-learning
PaperPDFCodeCodeCodeCodeCodeCode

Abstract

We propose a simple yet robust stochastic answer network (SAN) that simulates multi-step reasoning in machine reading comprehension. Compared to previous work such as ReasoNet which used reinforcement learning to determine the number of steps, the unique feature is the use of a kind of stochastic prediction dropout on the answer module (final layer) of the neural network during the training. We show that this simple trick improves robustness and achieves results competitive to the state-of-the-art on the Stanford Question Answering Dataset (SQuAD), the Adversarial SQuAD, and the Microsoft MAchine Reading COmprehension Dataset (MS MARCO).

Results

TaskDatasetMetricValueModel
Question AnsweringSQuAD1.1 devEM76.235SAN (single)
Question AnsweringSQuAD1.1 devF184.056SAN (single)
Question AnsweringSQuAD1.1EM79.608SAN (ensemble model)
Question AnsweringSQuAD1.1F186.496SAN (ensemble model)
Question AnsweringSQuAD1.1EM76.828SAN (single model)
Question AnsweringSQuAD1.1F184.396SAN (single model)
Question AnsweringSQuAD2.0EM71.316SAN (ensemble model)
Question AnsweringSQuAD2.0F173.704SAN (ensemble model)
Question AnsweringSQuAD2.0EM68.653SAN (single model)
Question AnsweringSQuAD2.0F171.439SAN (single model)

Related Papers

CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning2025-07-18From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17Spectral Bellman Method: Unifying Representation and Exploration in RL2025-07-17Aligning Humans and Robots via Reinforcement Learning from Implicit Human Feedback2025-07-17