TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Medical/Language Modelling/The Pile

Language Modelling on The Pile

Metric: Test perplexity (lower is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Test perplexity▲Extra DataPaperDate↕Code
1Larger Transformer 771M (fine-tuned)10NoNeed a Small Specialized Language Model? Plan Ea...2024-02-02-
2Hybrid H3 125M10.2NoHungry Hungry Hippos: Towards Language Modeling ...2022-12-28Code
3GPT-Neo 2.7B10.44NoKnowledge Unlearning for Mitigating Privacy Risk...2022-10-04Code
4Transformer 125M10.7NoHungry Hungry Hippos: Towards Language Modeling ...2022-12-28Code
5GPT-Neo 1.3B11.46NoKnowledge Unlearning for Mitigating Privacy Risk...2022-10-04Code
6Smaller Transformer 126M (fine-tuned)12NoNeed a Small Specialized Language Model? Plan Ea...2024-02-02-
7OPT 2.7B17.81NoKnowledge Unlearning for Mitigating Privacy Risk...2022-10-04Code
8GPT-Neo 125M17.83NoKnowledge Unlearning for Mitigating Privacy Risk...2022-10-04Code
9OPT 1.3B19.55NoKnowledge Unlearning for Mitigating Privacy Risk...2022-10-04Code
10Larger Transformer 771M (pre-trained)28.1NoNeed a Small Specialized Language Model? Plan Ea...2024-02-02-
11OPT 125M32.26NoKnowledge Unlearning for Mitigating Privacy Risk...2022-10-04Code
12Smaller Transformer 126M (pre-trained)33NoNeed a Small Specialized Language Model? Plan Ea...2024-02-02-