TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Paragraph-based Transformer Pre-training for Multi-Sentenc...

Paragraph-based Transformer Pre-training for Multi-Sentence Inference

Luca Di Liello, Siddhant Garg, Luca Soldaini, Alessandro Moschitti

2022-05-02NAACL 2022 7Question AnsweringAnswer SelectionFact Verification
PaperPDFCode(official)

Abstract

Inference tasks such as answer sentence selection (AS2) or fact verification are typically solved by fine-tuning transformer-based models as individual sentence-pair classifiers. Recent studies show that these tasks benefit from modeling dependencies across multiple candidate sentences jointly. In this paper, we first show that popular pre-trained transformers perform poorly when used for fine-tuning on multi-candidate inference tasks. We then propose a new pre-training objective that models the paragraph-level semantics across multiple input sentences. Our evaluation on three AS2 and one fact verification datasets demonstrates the superiority of our pre-training technique over the traditional ones for transformers used as joint models for multi-candidate inference tasks, as well as when used as cross-encoders for sentence-pair formulations of these tasks. Our code and pre-trained models are released at https://github.com/amazon-research/wqa-multi-sentence-inference .

Results

TaskDatasetMetricValueModel
Question AnsweringTrecQAMAP0.911RoBERTa-Base Joint + MSPP
Question AnsweringTrecQAMRR0.952RoBERTa-Base Joint + MSPP
Question AnsweringWikiQAMAP0.887RoBERTa-Base Joint MSPP
Question AnsweringWikiQAMRR0.9RoBERTa-Base Joint MSPP
Question AnsweringASNQMAP0.673RoBERTa-Base Joint MSPP
Question AnsweringASNQMRR0.737RoBERTa-Base Joint MSPP
Fact VerificationFEVERAccuracy75.36RoBERTa-Base Joint MSPP Flexible
Fact VerificationFEVERAccuracy74.39RoBERTa-Base Joint MSPP

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility2025-07-16Warehouse Spatial Question Answering with LLM Agent2025-07-14Evaluating Attribute Confusion in Fashion Text-to-Image Generation2025-07-09