TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Medical/Language Modelling/One Billion Word

Language Modelling on One Billion Word

Metric: Validation perplexity (lower is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Validation perplexity▲Extra DataPaperDate↕Code
1H-Transformer-1D Nr=16 (Large)20.25NoH-Transformer-1D: Fast One-Dimensional Hierarchi...2021-07-25Code
2Adaptive Input Very Large22.92NoAdaptive Input Representations for Neural Langua...2018-09-28Code
3Adaptive Input Large23.83NoAdaptive Input Representations for Neural Langua...2018-09-28Code
4H-Transformer-1D Nr=16 (Base)23.95NoH-Transformer-1D: Fast One-Dimensional Hierarchi...2021-07-25Code