TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Dense Representations of Phrases at Scale

Learning Dense Representations of Phrases at Scale

Jinhyuk Lee, Mujeen Sung, Jaewoo Kang, Danqi Chen

2020-12-23ACL 2021 5Reading ComprehensionQuestion AnsweringTransfer LearningSlot FillingOpen-Domain Question AnsweringQuestion GenerationRetrieval
PaperPDFCodeCodeCode(official)Code(official)

Abstract

Open-domain question answering can be reformulated as a phrase retrieval problem, without the need for processing documents on-demand during inference (Seo et al., 2019). However, current phrase retrieval models heavily depend on sparse representations and still underperform retriever-reader approaches. In this work, we show for the first time that we can learn dense representations of phrases alone that achieve much stronger performance in open-domain QA. We present an effective method to learn phrase representations from the supervision of reading comprehension tasks, coupled with novel negative sampling methods. We also propose a query-side fine-tuning strategy, which can support transfer learning and reduce the discrepancy between training and inference. On five popular open-domain QA datasets, our model DensePhrases improves over previous phrase retrieval models by 15%-25% absolute accuracy and matches the performance of state-of-the-art retriever-reader models. Our model is easy to parallelize due to pure dense representations and processes more than 10 questions per second on CPUs. Finally, we directly use our pre-indexed dense phrase representations for two slot filling tasks, showing the promise of utilizing DensePhrases as a dense knowledge base for downstream tasks.

Results

TaskDatasetMetricValueModel
Question AnsweringSQuAD1.1 devEM78.3DensePhrases
Question AnsweringSQuAD1.1 devF186.3DensePhrases
Question AnsweringNatural Questions (long)EM71.9DensePhrases
Question AnsweringNatural Questions (long)F179.6DensePhrases
Slot FillingKILT: T-RExAccuracy53.9DensePhrases
Slot FillingKILT: T-RExF161.74DensePhrases
Slot FillingKILT: T-RExKILT-AC27.84DensePhrases
Slot FillingKILT: T-RExKILT-F132.34DensePhrases
Slot FillingKILT: T-RExR-Prec37.62DensePhrases
Slot FillingKILT: T-RExRecall@540.07DensePhrases
Slot FillingKILT: Zero Shot REAccuracy47.42DensePhrases
Slot FillingKILT: Zero Shot REF154.75DensePhrases
Slot FillingKILT: Zero Shot REKILT-AC41.34DensePhrases
Slot FillingKILT: Zero Shot REKILT-F146.79DensePhrases
Slot FillingKILT: Zero Shot RER-Prec57.43DensePhrases
Slot FillingKILT: Zero Shot RERecall@560.47DensePhrases

Related Papers

RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17