TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Natural Language Processing/Reading Comprehension/MuSeRC

Reading Comprehension on MuSeRC

Metric: Average F1 (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Average F1▼Extra DataPaperDate↕Code
1Golden Transformer0.941No---
2MT5 Large0.844NomT5: A massively multilingual pre-trained text-t...2020-10-22Code
3ruRoberta-large finetune0.83No---
4ruT5-large-finetune0.815No---
5Human Benchmark0.806NoRussianSuperGLUE: A Russian Language Understandi...2020-10-29Code
6ruT5-base-finetune0.769No---
7ruBert-large finetune0.76No---
8ruBert-base finetune0.742No---
9RuGPT3XL few-shot0.74No---
10RuGPT3Large0.729No---
11RuBERT plain0.711No---
12RuGPT3Medium0.706No---
13RuBERT conversational0.687No---
14YaLM 1.0B few-shot0.673No---
15heuristic majority0.671NoUnreasonable Effectiveness of Rule-Based Heurist...2021-05-03-
16RuGPT3Small0.653No---
17SBERT_Large0.646No---
18SBERT_Large_mt_ru_finetuning0.642No---
19Multilingual Bert0.639No---
20Baseline TF-IDF1.10.587NoRussianSuperGLUE: A Russian Language Understandi...2020-10-29Code
21Random weighted0.45NoUnreasonable Effectiveness of Rule-Based Heurist...2021-05-03-
22majority_class0No---