TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Natural Language Inference/WNLI

Natural Language Inference on WNLI

Metric: Accuracy (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Accuracy▼Extra DataPaperDate↕Code
1Turing NLR v5 XXL 5.4B (fine-tuned)95.9NoToward Efficient Language Model Pretraining and ...2022-12-04-
2DeBERTa94.5NoDeBERTa: Decoding-enhanced BERT with Disentangle...2020-06-05Code
3T5-XXL 11B93.2NoExploring the Limits of Transfer Learning with a...2019-10-23Code
4XLNet92.5NoXLNet: Generalized Autoregressive Pretraining fo...2019-06-19Code
5ALBERT91.8NoALBERT: A Lite BERT for Self-supervised Learning...2019-09-26Code
6T5-XL 3B89.7NoExploring the Limits of Transfer Learning with a...2019-10-23Code
7StructBERTRoBERTa ensemble89.7NoStructBERT: Incorporating Language Structures in...2019-08-13-
8HNNensemble89NoA Hybrid Neural Network Model for Commonsense Re...2019-07-27Code
9RoBERTa (ensemble)89NoRoBERTa: A Robustly Optimized BERT Pretraining A...2019-07-26Code
10T5-Large 770M85.6NoExploring the Limits of Transfer Learning with a...2019-10-23Code
11HNN83.6NoA Hybrid Neural Network Model for Commonsense Re...2019-07-27Code
12T5-Base 220M78.8NoExploring the Limits of Transfer Learning with a...2019-10-23Code
13BERTwiki 340M (fine-tuned on WSCR)74.7NoA Surprisingly Robust Trick for Winograd Schema ...2019-05-15Code
14FLAN 137B (zero-shot)74.6NoFinetuned Language Models Are Zero-Shot Learners2021-09-03Code
15BERT-large 340M (fine-tuned on WSCR)71.9NoA Surprisingly Robust Trick for Winograd Schema ...2019-05-15Code
16BERT-base 110M (fine-tuned on WSCR)70.5NoA Surprisingly Robust Trick for Winograd Schema ...2019-05-15Code
17FLAN 137B (few-shot, k=4)70.4NoFinetuned Language Models Are Zero-Shot Learners2021-09-03Code
18T5-Small 60M69.2NoExploring the Limits of Transfer Learning with a...2019-10-23Code
19ERNIE 2.0 Large67.8NoERNIE 2.0: A Continual Pre-training Framework fo...2019-07-29Code
20SqueezeBERT65.1NoSqueezeBERT: What can computer vision teach NLP ...2020-06-19Code
21BERT-large 340M65.1NoBERT: Pre-training of Deep Bidirectional Transfo...2018-10-11Code
22RWKV-4-Raven-14B49.3NoRWKV: Reinventing RNNs for the Transformer Era2023-05-22Code
23DistilBERT 66M44.4NoDistilBERT, a distilled version of BERT: smaller...2019-10-02Code