TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Question Answering/SQuAD2.0 dev

Question Answering on SQuAD2.0 dev

Metric: EM (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕EM▼Extra DataPaperDate↕Code
1XLNet (single model)87.9NoXLNet: Generalized Autoregressive Pretraining fo...2019-06-19Code
2XLNet+DSC87.65NoDice Loss for Data-imbalanced NLP Tasks2019-11-07Code
3RoBERTa (no data aug)86.5YesRoBERTa: A Robustly Optimized BERT Pretraining A...2019-07-26Code
4ALBERT xxlarge85.1NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
5SG-Net85.1NoSG-Net: Syntax-Guided Machine Reading Comprehens...2019-08-14Code
6ALBERT xlarge83.1NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
7SemBERT large80.9NoSemantics-aware BERT for Language Understanding2019-09-05Code
8ALBERT large79NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
9ALBERT base76.1NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
10RMR + ELMo (Model-III)72.3NoRead + Verify: Machine Reading Comprehension wit...2018-08-17-
11U-Net70.3NoU-Net: Machine Reading Comprehension with Unansw...2018-10-12Code
12TinyBERT-6 67M69.9NoTinyBERT: Distilling BERT for Natural Language U...2019-09-23Code