TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Datasets/GLUE

GLUE

General Language Understanding Evaluation benchmark

TextsCustom (various)Introduced 2019-01-01

General Language Understanding Evaluation (GLUE) benchmark is a collection of nine natural language understanding tasks, including single-sentence tasks CoLA and SST-2, similarity and paraphrasing tasks MRPC, STS-B and QQP, and natural language inference tasks MNLI, QNLI, RTE and WNLI.

Source: Align, Mask and Select: A Simple Method for Incorporating Commonsense Knowledge into Language Representation Models Image Source: https://gluebenchmark.com/

Benchmarks

Natural Language Understanding/Average

Related Benchmarks

GLUE COLA/Classification/Matthews CorrelationGLUE COLA/Text Classification/Matthews CorrelationGLUE MRPC/Classification/AccuracyGLUE MRPC/Classification/F1GLUE MRPC/Text Classification/AccuracyGLUE MRPC/Text Classification/F1GLUE QQP/Few-Shot Learning/F1-scoreGLUE QQP/Meta-Learning/F1-scoreGLUE RTE/Classification/AccuracyGLUE RTE/Text Classification/AccuracyGLUE SST2/Classification/AccuracyGLUE SST2/Text Classification/AccuracyGLUE STSB/Classification/SpearmanrGLUE STSB/Text Classification/Spearmanr

Statistics

Papers
3,197
Benchmarks
1

Links

Homepage

Tasks

ClassificationData-free Knowledge DistillationFew-Shot LearningLinguistic AcceptabilityModel CompressionNatural Language InferenceNatural Language UnderstandingQQPSemantic Textual SimilaritySemantic Textual Similarity within Bi-EncoderSentence-EmbeddingStochastic OptimizationText ClassificationText Generation