TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SpanBERT: Improving Pre-training by Representing and Predi...

SpanBERT: Improving Pre-training by Representing and Predicting Spans

Mandar Joshi, Danqi Chen, Yinhan Liu, Daniel S. Weld, Luke Zettlemoyer, Omer Levy

2019-07-24TACL 2020 1Question AnsweringRelation ExtractionParaphrase IdentificationSentiment AnalysisCoreference ResolutionNatural Language InferenceSemantic Textual SimilarityLinguistic AcceptabilityOpen-Domain Question AnsweringRelation Classification
PaperPDFCodeCodeCode(official)CodeCodeCode

Abstract

We present SpanBERT, a pre-training method that is designed to better represent and predict spans of text. Our approach extends BERT by (1) masking contiguous random spans, rather than random tokens, and (2) training the span boundary representations to predict the entire content of the masked span, without relying on the individual token representations within it. SpanBERT consistently outperforms BERT and our better-tuned baselines, with substantial gains on span selection tasks such as question answering and coreference resolution. In particular, with the same training data and model size as BERT-large, our single model obtains 94.6% and 88.7% F1 on SQuAD 1.1 and 2.0, respectively. We also achieve a new state of the art on the OntoNotes coreference resolution task (79.6\% F1), strong performance on the TACRED relation extraction benchmark, and even show gains on GLUE.

Results

TaskDatasetMetricValueModel
Relation ExtractionTACREDF170.8SpanBERT-large
Relation ExtractionRe-TACREDF185.3SpanBERT
Relation ExtractionTACREDF170.8SpanBERT
Relation ClassificationTACREDF170.8SpanBERT
Question AnsweringNewsQAF173.6SpanBERT
Question AnsweringNaturalQAF182.5SpanBERT
Question AnsweringSQuAD1.1EM88.8SpanBERT (single model)
Question AnsweringSQuAD1.1F194.6SpanBERT (single model)
Question AnsweringTriviaQAF183.6SpanBERT
Question AnsweringSQuAD2.0 devF186.8SpanBERT
Question AnsweringSQuAD2.0EM85.7SpanBERT
Question AnsweringSQuAD2.0F188.7SpanBERT
Question AnsweringSearchQAF184.8SpanBERT
Natural Language InferenceMultiNLIMatched88.1SpanBERT
Semantic Textual SimilaritySTS BenchmarkPearson Correlation0.899SpanBERT
Semantic Textual SimilarityQuora Question PairsAccuracy89.5SpanBERT
Semantic Textual SimilarityQuora Question PairsF171.9SpanBERT
Sentiment AnalysisSST-2 Binary classificationAccuracy94.8SpanBERT
Coreference ResolutionOntoNotesF179.6SpanBERT
Paraphrase IdentificationQuora Question PairsAccuracy89.5SpanBERT
Paraphrase IdentificationQuora Question PairsF171.9SpanBERT
Open-Domain Question AnsweringSearchQAF184.8SpanBERT

Related Papers

From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17Enter the Mind Palace: Reasoning and Planning for Long-term Active Embodied Question Answering2025-07-17Vision-and-Language Training Helps Deploy Taxonomic Knowledge but Does Not Fundamentally Alter It2025-07-17City-VLM: Towards Multidomain Perception Scene Understanding via Multimodal Incomplete Learning2025-07-17AdaptiSent: Context-Aware Adaptive Attention for Multimodal Aspect-Based Sentiment Analysis2025-07-17SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts2025-07-17Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility2025-07-16