TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Text Understanding with the Attention Sum Reader Network

Text Understanding with the Attention Sum Reader Network

Rudolf Kadlec, Martin Schmid, Ondrej Bajgar, Jan Kleindienst

2016-03-04ACL 2016 8Reading ComprehensionQuestion AnsweringOpen-Domain Question AnsweringMachine Reading Comprehension
PaperPDFCode(official)Code

Abstract

Several large cloze-style context-question-answer datasets have been introduced recently: the CNN and Daily Mail news data and the Children's Book Test. Thanks to the size of these datasets, the associated text comprehension task is well suited for deep-learning techniques that currently seem to outperform all alternative approaches. We present a new, simple model that uses attention to directly pick the answer from the context as opposed to computing the answer using a blended representation of words in the document as is usual in similar models. This makes the model particularly suitable for question-answering problems where the answer is a single word from the document. Ensemble of our models sets new state of the art on all evaluated datasets.

Results

TaskDatasetMetricValueModel
Question AnsweringCNN / Daily MailCNN75.4AS Reader (ensemble model)
Question AnsweringCNN / Daily MailDaily Mail77.7AS Reader (ensemble model)
Question AnsweringCNN / Daily MailCNN69.5AS Reader (single model)
Question AnsweringCNN / Daily MailDaily Mail73.9AS Reader (single model)
Question AnsweringSearchQAN-gram F122.8ASR
Question AnsweringSearchQAUnigram Acc41.3ASR
Open-Domain Question AnsweringSearchQAN-gram F122.8ASR
Open-Domain Question AnsweringSearchQAUnigram Acc41.3ASR

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility2025-07-16Warehouse Spatial Question Answering with LLM Agent2025-07-14Evaluating Attribute Confusion in Fashion Text-to-Image Generation2025-07-09