TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Question Answering/SQuAD2.0 dev

Question Answering on SQuAD2.0 dev

Metric: F1 (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕F1▼Extra DataPaperDate↕Code
1XLNet (single model)90.6NoXLNet: Generalized Autoregressive Pretraining fo...2019-06-19Code
2XLNet+DSC89.51NoDice Loss for Data-imbalanced NLP Tasks2019-11-07Code
3RoBERTa (no data aug)89.4YesRoBERTa: A Robustly Optimized BERT Pretraining A...2019-07-26Code
4ALBERT xxlarge88.1NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
5SG-Net87.9NoSG-Net: Syntax-Guided Machine Reading Comprehens...2019-08-14Code
6SpanBERT86.8NoSpanBERT: Improving Pre-training by Representing...2019-07-24Code
7ALBERT xlarge85.9NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
8SemBERT large83.6NoSemantics-aware BERT for Language Understanding2019-09-05Code
9ALBERT large82.1NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
10ALBERT base79.1NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
11RMR + ELMo (Model-III)74.8NoRead + Verify: Machine Reading Comprehension wit...2018-08-17-
12U-Net74NoU-Net: Machine Reading Comprehension with Unansw...2018-10-12Code
13TinyBERT-6 67M73.4NoTinyBERT: Distilling BERT for Natural Language U...2019-09-23Code