TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

SotA/Methodology/Transfer Learning/MGSM

Transfer Learning on MGSM

Metric: Average (%) (higher is better)

LeaderboardDataset
Loading chart...

Results

Submit a result
#Model↕Average (%)▼AugmentationsPaperDate↕Code
1PaLM 2 (few-shot, k=8, SC)87NoPaLM 2 Technical Report2023-05-17Code
2PaLM 2 (8-shot, CoT)72.2NoPaLM 2 Technical Report2023-05-17Code
3Flan-PaLM 540B (8-shot, fine-tuned, CoT + SC)72NoScaling Instruction-Finetuned Language Models2022-10-20Code
4Flan-U-PaLM 540B (CoT)60.4NoScaling Instruction-Finetuned Language Models2022-10-20Code
5Flan-PaLM 540B (8-shot, fine-tuned, CoT)57NoScaling Instruction-Finetuned Language Models2022-10-20Code
6PaLM 540B55NoPaLM: Scaling Language Modeling with Pathways2022-04-05Code
7U-PaLM 540B (CoT)49.9NoTranscending Scaling Laws with 0.1% Extra Compute2022-10-20-
8text-davinci-00336NoScaling Instruction-Finetuned Language Models2022-10-20Code
9code-davinci-00235NoScaling Instruction-Finetuned Language Models2022-10-20Code
10text-davinci-00223.7NoScaling Instruction-Finetuned Language Models2022-10-20Code
11Flan-PaLM 540B (8-shot, fine-tuned)21.2NoScaling Instruction-Finetuned Language Models2022-10-20Code
12GPT-3 Davinci 175B5.7NoScaling Instruction-Finetuned Language Models2022-10-20Code