TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Bidirectional Attention Flow for Machine Comprehension

Bidirectional Attention Flow for Machine Comprehension

Minjoon Seo, Aniruddha Kembhavi, Ali Farhadi, Hannaneh Hajishirzi

2016-11-05Reading ComprehensionQuestion AnsweringCloze TestNavigateOpen-Domain Question Answering
PaperPDFCodeCodeCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Machine comprehension (MC), answering a query about a given context paragraph, requires modeling complex interactions between the context and the query. Recently, attention mechanisms have been successfully extended to MC. Typically these methods use attention to focus on a small portion of the context and summarize it with a fixed-size vector, couple attentions temporally, and/or often form a uni-directional attention. In this paper we introduce the Bi-Directional Attention Flow (BIDAF) network, a multi-stage hierarchical process that represents the context at different levels of granularity and uses bi-directional attention flow mechanism to obtain a query-aware context representation without early summarization. Our experimental evaluations show that our model achieves the state-of-the-art results in Stanford Question Answering Dataset (SQuAD) and CNN/DailyMail cloze test.

Results

TaskDatasetMetricValueModel
Reading ComprehensionAdversarialQAOverall: F128.5BiDAF
Question AnsweringNarrativeQABLEU-133.45BiDAF
Question AnsweringNarrativeQABLEU-415.69BiDAF
Question AnsweringNarrativeQAMETEOR15.68BiDAF
Question AnsweringNarrativeQARouge-L36.74BiDAF
Question AnsweringSQuAD1.1 devEM67.7BIDAF (single)
Question AnsweringSQuAD1.1 devF177.3BIDAF (single)
Question AnsweringMS MARCOBLEU-110.64BiDaF Baseline
Question AnsweringMS MARCORouge-L23.96BiDaF Baseline
Question AnsweringSQuAD1.1EM73.744BiDAF (ensemble)
Question AnsweringSQuAD1.1F181.525BiDAF (ensemble)
Question AnsweringSQuAD1.1EM67.974BiDAF (single model)
Question AnsweringSQuAD1.1F177.323BiDAF (single model)
Question AnsweringCNN / Daily MailCNN76.9BiDAF
Question AnsweringCNN / Daily MailDaily Mail79.6BiDAF
Question AnsweringQuasarEM (Quasar-T)25.9BiDAF
Question AnsweringQuasarF1 (Quasar-T)28.5BiDAF
Open-Domain Question AnsweringQuasarEM (Quasar-T)25.9BiDAF
Open-Domain Question AnsweringQuasarF1 (Quasar-T)28.5BiDAF

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility2025-07-16Vision-based Perception for Autonomous Vehicles in Obstacle Avoidance Scenarios2025-07-16CogDDN: A Cognitive Demand-Driven Navigation with Decision Optimization and Dual-Process Thinking2025-07-15