Tasks
SotA
Datasets
Papers
Methods
Submit
About
SotA
/
Natural Language Processing
/
Question Answering
/
StoryCloze
Question Answering on StoryCloze
Metric: Accuracy (higher is better)
Leaderboard
Dataset
Loading chart...
Results
Submit a result
Export CSV
#
Model
↕
Accuracy
▼
Extra Data
Paper
Date
↕
Code
1
BLOOMZ
96.3
No
Crosslingual Generalization through Multitask Fi...
2022-11-03
Code
2
Flipped-3B
95.88
No
Guess the Instruction! Flipped Learning Makes La...
2022-10-06
Code
3
FLAN 137B (few-shot, k=10)
94.7
No
Finetuned Language Models Are Zero-Shot Learners
2021-09-03
Code
4
T0-3B (CoT fine-tuned)
94.5
No
The CoT Collection: Improving Zero-shot and Few-...
2023-05-23
Code
5
KiC-770M
94.4
No
Knowledge-in-Context: Towards Knowledgeable Semi...
2022-10-28
-
6
FLAN 137B (zero-shot)
93.4
No
Finetuned Language Models Are Zero-Shot Learners
2021-09-03
Code
7
Reading Strategies Model
88.3
No
Improving Machine Reading Comprehension with Gen...
2018-10-31
Code
8
Finetuned Transformer LM
86.5
No
-
-
Code
9
RoE-3B
86.33
No
Exploring the Benefits of Training Expert Langua...
2023-02-07
Code
10
OPT-175B
79.82
No
SparseGPT: Massive Language Models Can Be Accura...
2023-01-02
Code
11
SparseGPT (175B, 50% Sparsity)
78.87
No
SparseGPT: Massive Language Models Can Be Accura...
2023-01-02
Code
12
Memory chains and semantic supervision
78.7
No
-
-
Code
13
Hidden Coherence Model
77.6
No
-
-
-
14
SparseGPT (175B, 4:8 Sparsity)
77.02
No
SparseGPT: Massive Language Models Can Be Accura...
2023-01-02
Code
15
val-LS-skip
76.5
No
A Simple and Effective Approach to the Story Clo...
2018-03-15
-
16
SparseGPT (175B, 2:4 Sparsity)
76.19
No
SparseGPT: Massive Language Models Can Be Accura...
2023-01-02
Code
17
sMLP – deterministic 9.4B (0-shot)
74.7
No
Efficient Language Modeling with Sparse all-MLP
2022-03-14
-
18
Switch Transformer 9B
73.3
No
Efficient Language Modeling with Sparse all-MLP
2022-03-14
-
19
GPT-3 Large 760M (zero-shot)
72.4
No
Language Models are Few-Shot Learners
2020-05-28
Code
20
Gshard 9B
67.9
No
Efficient Language Modeling with Sparse all-MLP
2022-03-14
-
21
HASH Layers 10B (0-shot)
64.7
No
Efficient Language Modeling with Sparse all-MLP
2022-03-14
-
22
Base Layers 10B (0-shot)
61.4
No
Efficient Language Modeling with Sparse all-MLP
2022-03-14
-
23
OPT-175B (50% Sparsity)
47.1
No
SparseGPT: Massive Language Models Can Be Accura...
2023-01-02
Code