TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Pre-training Transformer Models with Sentence-Level Object...

Pre-training Transformer Models with Sentence-Level Objectives for Answer Sentence Selection

Luca Di Liello, Siddhant Garg, Luca Soldaini, Alessandro Moschitti

2022-05-20Question AnsweringAnswer Selection
PaperPDF

Abstract

An important task for designing QA systems is answer sentence selection (AS2): selecting the sentence containing (or constituting) the answer to a question from a set of retrieved relevant documents. In this paper, we propose three novel sentence-level transformer pre-training objectives that incorporate paragraph-level semantics within and across documents, to improve the performance of transformers for AS2, and mitigate the requirement of large labeled datasets. Specifically, the model is tasked to predict whether: (i) two sentences are extracted from the same paragraph, (ii) a given sentence is extracted from a given paragraph, and (iii) two paragraphs are extracted from the same document. Our experiments on three public and one industrial AS2 datasets demonstrate the empirical superiority of our pre-trained transformers over baseline models such as RoBERTa and ELECTRA for AS2.

Results

TaskDatasetMetricValueModel
Question AnsweringTrecQAMAP0.923DeBERTa-V3-Large + SSP
Question AnsweringTrecQAMRR0.946DeBERTa-V3-Large + SSP
Question AnsweringTrecQAMAP0.903RoBERTa-Base + PSD
Question AnsweringTrecQAMRR0.951RoBERTa-Base + PSD
Question AnsweringWikiQAMAP0.909DeBERTa-V3-Large + ALL
Question AnsweringWikiQAMRR0.92DeBERTa-V3-Large + ALL
Question AnsweringWikiQAMAP0.901DeBERTa-Large + SSP
Question AnsweringWikiQAMRR0.914DeBERTa-Large + SSP
Question AnsweringWikiQAMAP0.887RoBERTa-Base + SSP
Question AnsweringWikiQAMRR0.899RoBERTa-Base + SSP
Question AnsweringASNQMAP0.743DeBERTa-V3-Large + SSP
Question AnsweringASNQMRR0.8DeBERTa-V3-Large + SSP
Question AnsweringASNQMAP0.697ELECTRA-Base + SSP
Question AnsweringASNQMRR0.757ELECTRA-Base + SSP

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility2025-07-16Warehouse Spatial Question Answering with LLM Agent2025-07-14Evaluating Attribute Confusion in Fashion Text-to-Image Generation2025-07-09