TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Common Sense Reasoning/ReCoRD

Common Sense Reasoning on ReCoRD

Metric: EM (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕EM▼Extra DataPaperDate↕Code
1Turing NLR v5 XXL 5.4B (fine-tuned)95.9NoToward Efficient Language Model Pretraining and ...2022-12-04-
2ST-MoE-32B 269B (fine-tuned)95.1NoST-MoE: Designing Stable and Transferable Sparse...2022-02-17Code
3DeBERTa-1.5B94.1NoDeBERTa: Decoding-enhanced BERT with Disentangle...2020-06-05Code
4PaLM 540B (finetuned) 94NoPaLM: Scaling Language Modeling with Pathways2022-04-05Code
5Vega v2 6B (fine-tuned)93.9NoToward Efficient Language Model Pretraining and ...2022-12-04-
6T5-XXL 11B (fine-tuned)93.4NoExploring the Limits of Transfer Learning with a...2019-10-23Code
7GESA 500M91.7NoIntegrating a Heterogeneous Graph with Entity-aw...2023-07-19-
8LUKE-Graph91.2NoLUKE-Graph: A Transformer-based Approach with Ga...2023-03-12-
9LUKE (single model)90.64No---
10LUKE 483M90.6NoLUKE: Deep Contextualized Entity Representations...2020-10-02Code
11KELM (finetuning RoBERTa-large based single model)89.1NoKELM: Knowledge Enhanced Pre-Trained Language Re...2021-09-09Code
12ST-MoE-L 4.1B (fine-tuned)88.9NoST-MoE: Designing Stable and Transferable Sparse...2022-02-17Code
13FLAN 137B (prompt-tuned)85.1NoFinetuned Language Models Are Zero-Shot Learners2021-09-03Code
14XLNet + MTL + Verifier (ensemble)83.09No---
15GPT-3 Large 760M (0-shot)82.1NoLanguage Models are Few-Shot Learners2020-05-28Code
16CSRLM (single model)81.78No---
17XLNet + Verifier81.5No---
18XLNet + MTL + Verifier (single model)81.46No---
19Switch Transformer 9B79.9NoEfficient Language Modeling with Sparse all-MLP2022-03-14-
20{SKG-NET} (single model)79.48No---
21KELM (finetuning BERT-large based single model)76.2NoKELM: Knowledge Enhanced Pre-Trained Language Re...2021-09-09Code
22sMLP – deterministic 9.4B (0-shot)73.4NoEfficient Language Modeling with Sparse all-MLP2022-03-14-
23FLAN 137B (zero-shot)72.5NoFinetuned Language Models Are Zero-Shot Learners2021-09-03Code
24Gshard 9B72.4NoEfficient Language Modeling with Sparse all-MLP2022-03-14-
25SKG-BERT (single model)72.24No---
26KT-NET (single model)71.6No---
27DCReader+BERT (single model)69.49No---
28HASH Layers 10B (0-shot)67.2NoEfficient Language Modeling with Sparse all-MLP2022-03-14-
29GraphBert (single)60.8No---
30Base Layers 10B (0-shot)60.7NoEfficient Language Modeling with Sparse all-MLP2022-03-14-
31GraphBert-WordNet (single)59.86No---
32GraphBert-NELL (single)59.41No---
33BERT-Base (single model)54.04NoBERT: Pre-training of Deep Bidirectional Transfo...2018-10-11Code
34DocQA + ELMo45.4NoReCoRD: Bridging the Gap between Human and Machi...2018-10-30-
35N-Grammer 343M28.9NoN-Grammer: Augmenting Transformers with latent n...2022-07-13Code