TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/RuCoLA: Russian Corpus of Linguistic Acceptability

RuCoLA: Russian Corpus of Linguistic Acceptability

Vladislav Mikhailov, Tatiana Shamardina, Max Ryabinin, Alena Pestova, Ivan Smurov, Ekaterina Artemova

2022-10-23Text GenerationLinguistic Acceptability
PaperPDFCode(official)

Abstract

Linguistic acceptability (LA) attracts the attention of the research community due to its many uses, such as testing the grammatical knowledge of language models and filtering implausible texts with acceptability classifiers. However, the application scope of LA in languages other than English is limited due to the lack of high-quality resources. To this end, we introduce the Russian Corpus of Linguistic Acceptability (RuCoLA), built from the ground up under the well-established binary LA approach. RuCoLA consists of $9.8$k in-domain sentences from linguistic publications and $3.6$k out-of-domain sentences produced by generative models. The out-of-domain set is created to facilitate the practical use of acceptability for improving language generation. Our paper describes the data collection protocol and presents a fine-grained analysis of acceptability classification experiments with a range of baseline approaches. In particular, we demonstrate that the most widely used language models still fall behind humans by a large margin, especially when detecting morphological and semantic errors. We release RuCoLA, the code of experiments, and a public leaderboard (rucola-benchmark.com) to assess the linguistic competence of language models for Russian.

Results

TaskDatasetMetricValueModel
Linguistic AcceptabilityRuCoLAAccuracy79.34ruRoBERTa
Linguistic AcceptabilityRuCoLAMCC0.53ruRoBERTa
Linguistic AcceptabilityRuCoLAAccuracy75.06RemBERT
Linguistic AcceptabilityRuCoLAMCC0.44RemBERT
Linguistic AcceptabilityRuCoLAAccuracy74.3ruBERT
Linguistic AcceptabilityRuCoLAMCC0.42ruBERT
Linguistic AcceptabilityRuCoLAAccuracy53.82ruGPT-3
Linguistic AcceptabilityRuCoLAMCC0.3ruGPT-3
Linguistic AcceptabilityRuCoLAAccuracy68.41ruT5
Linguistic AcceptabilityRuCoLAMCC0.25ruT5
Linguistic AcceptabilityRuCoLAMCC0.15mBERT
Linguistic AcceptabilityRuCoLAAccuracy61.13XLM-R
Linguistic AcceptabilityRuCoLAMCC0.13XLM-R
Linguistic AcceptabilityCoLAMCC0.6RemBERT
Linguistic AcceptabilityItaCoLAMCC0.52XLM-R
Linguistic AcceptabilityItaCoLAMCC0.36mBERT

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs2025-07-15Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Exploiting Leaderboards for Large-Scale Distribution of Malicious Models2025-07-11CLI-RAG: A Retrieval-Augmented Framework for Clinically Structured and Context Aware Text Generation with LLMs2025-07-09FIFA: Unified Faithfulness Evaluation Framework for Text-to-Video and Video-to-Text Generation2025-07-09