TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Reading Wikipedia to Answer Open-Domain Questions

Reading Wikipedia to Answer Open-Domain Questions

Danqi Chen, Adam Fisch, Jason Weston, Antoine Bordes

2017-03-31ACL 2017 7Reading ComprehensionQuestion AnsweringOpen-Domain Question AnsweringRetrieval
PaperPDFCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCode

Abstract

This paper proposes to tackle open- domain question answering using Wikipedia as the unique knowledge source: the answer to any factoid question is a text span in a Wikipedia article. This task of machine reading at scale combines the challenges of document retrieval (finding the relevant articles) with that of machine comprehension of text (identifying the answer spans from those articles). Our approach combines a search component based on bigram hashing and TF-IDF matching with a multi-layer recurrent neural network model trained to detect answers in Wikipedia paragraphs. Our experiments on multiple existing QA datasets indicate that (1) both modules are highly competitive with respect to existing counterparts and (2) multitask learning using distant supervision on their combination is an effective complete system on this challenging task.

Results

TaskDatasetMetricValueModel
Question AnsweringSQuAD1.1 devEM69.5DrQA (Document Reader only)
Question AnsweringSQuAD1.1 devF178.8DrQA (Document Reader only)
Question AnsweringQuasart-TEM37.7DrQA
Question AnsweringSQuAD1.1EM70.733Document Reader (single model)
Question AnsweringSQuAD1.1F179.353Document Reader (single model)
Question AnsweringNatural Questions (long)F146.1DrQA
Question AnsweringSearchQAEM41.9DrQA
Question AnsweringSQuAD1.1EM70DrQA
Open-Domain Question AnsweringSearchQAEM41.9DrQA
Open-Domain Question AnsweringSQuAD1.1EM70DrQA

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16