TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Re2G: Retrieve, Rerank, Generate

Re2G: Retrieve, Rerank, Generate

Michael Glass, Gaetano Rossiello, Md Faisal Mahbub Chowdhury, Ankita Rajaram Naik, Pengshan Cai, Alfio Gliozzo

2022-07-13NAACL 2022 7Question AnsweringRerankingZero-shot Slot FillingFact CheckingSlot FillingOpen-Domain Question AnsweringRetrievalKnowledge DistillationOpen-Domain DialogFact VerificationRAG
PaperPDFCode(official)

Abstract

As demonstrated by GPT-3 and T5, transformers grow in capability as parameter spaces become larger and larger. However, for tasks that require a large amount of knowledge, non-parametric memory allows models to grow dramatically with a sub-linear increase in computational cost and GPU memory requirements. Recent models such as RAG and REALM have introduced retrieval into conditional generation. These models incorporate neural initial retrieval from a corpus of passages. We build on this line of research, proposing Re2G, which combines both neural initial retrieval and reranking into a BART-based sequence-to-sequence generation. Our reranking approach also permits merging retrieval results from sources with incomparable scores, enabling an ensemble of BM25 and neural initial retrieval. To train our system end-to-end, we introduce a novel variation of knowledge distillation to train the initial retrieval, reranker, and generation using only ground truth on the target sequence output. We find large gains in four diverse tasks: zero-shot slot filling, question answering, fact-checking, and dialog, with relative gains of 9% to 34% over the previous state-of-the-art on the KILT leaderboard. We make our code available as open source at https://github.com/IBM/kgi-slot-filling/tree/re2g.

Results

TaskDatasetMetricValueModel
Question AnsweringKILT: TriviaQAEM76.27Re2G
Question AnsweringKILT: TriviaQAF181.4Re2G
Question AnsweringKILT: TriviaQAKILT-EM57.91Re2G
Question AnsweringKILT: TriviaQAKILT-F161.78Re2G
Question AnsweringKILT: TriviaQAR-Prec72.68Re2G
Question AnsweringKILT: TriviaQARecall@574.23Re2G
Question AnsweringKILT: Natural QuestionsEM51.73Re2G
Question AnsweringKILT: Natural QuestionsF160.97Re2G
Question AnsweringKILT: Natural QuestionsKILT-EM43.56Re2G
Question AnsweringKILT: Natural QuestionsKILT-F149.8Re2G
Question AnsweringKILT: Natural QuestionsR-Prec70.78Re2G
Question AnsweringKILT: Natural QuestionsRecall@576.63Re2G
Slot FillingKILT: T-RExAccuracy87.68Re2G
Slot FillingKILT: T-RExF189.93Re2G
Slot FillingKILT: T-RExKILT-AC75.84Re2G
Slot FillingKILT: T-RExKILT-F177.05Re2G
Slot FillingKILT: T-RExR-Prec80.7Re2G
Slot FillingKILT: T-RExRecall@589Re2G
Fact VerificationKILT: FEVERAccuracy89.55Re2G
Fact VerificationKILT: FEVERKILT-AC78.53Re2G
Fact VerificationKILT: FEVERR-Prec88.92Re2G
Fact VerificationKILT: FEVERRecall@592.52Re2G
Open-Domain Question AnsweringKILT: TriviaQAEM76.27Re2G
Open-Domain Question AnsweringKILT: TriviaQAF181.4Re2G
Open-Domain Question AnsweringKILT: TriviaQAKILT-EM57.91Re2G
Open-Domain Question AnsweringKILT: TriviaQAKILT-F161.78Re2G
Open-Domain Question AnsweringKILT: TriviaQAR-Prec72.68Re2G
Open-Domain Question AnsweringKILT: TriviaQARecall@574.23Re2G
Open-Domain Question AnsweringKILT: Natural QuestionsEM51.73Re2G
Open-Domain Question AnsweringKILT: Natural QuestionsF160.97Re2G
Open-Domain Question AnsweringKILT: Natural QuestionsKILT-EM43.56Re2G
Open-Domain Question AnsweringKILT: Natural QuestionsKILT-F149.8Re2G
Open-Domain Question AnsweringKILT: Natural QuestionsR-Prec70.78Re2G
Open-Domain Question AnsweringKILT: Natural QuestionsRecall@576.63Re2G
Open-Domain DialogKILT: Wizard of WikipediaF118.9Re2G
Open-Domain DialogKILT: Wizard of WikipediaKILT-F112.98Re2G
Open-Domain DialogKILT: Wizard of WikipediaKILT-RL11.39Re2G
Open-Domain DialogKILT: Wizard of WikipediaR-Prec60.1Re2G
Open-Domain DialogKILT: Wizard of WikipediaROUGE-L16.76Re2G
Open-Domain DialogKILT: Wizard of WikipediaRecall@579.98Re2G

Related Papers

PiMRef: Detecting and Explaining Ever-evolving Spear Phishing Emails with Knowledge Base Invariants2025-07-21Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17