TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/First Train to Generate, then Generate to Train: UnitedSyn...

First Train to Generate, then Generate to Train: UnitedSynT5 for Few-Shot NLI

Sourav Banerjee, Anush Mahajan, Ayushi Agarwal, Eishkaran Singh

2024-12-12Few-Shot LearningNatural Language InferenceNatural Language UnderstandingData AugmentationFew-Shot NLI
PaperPDF

Abstract

Natural Language Inference (NLI) tasks require identifying the relationship between sentence pairs, typically classified as entailment, contradiction, or neutrality. While the current state-of-the-art (SOTA) model, Entailment Few-Shot Learning (EFL), achieves a 93.1% accuracy on the Stanford Natural Language Inference (SNLI) dataset, further advancements are constrained by the dataset's limitations. To address this, we propose a novel approach leveraging synthetic data augmentation to enhance dataset diversity and complexity. We present UnitedSynT5, an advanced extension of EFL that leverages a T5-based generator to synthesize additional premise-hypothesis pairs, which are rigorously cleaned and integrated into the training data. These augmented examples are processed within the EFL framework, embedding labels directly into hypotheses for consistency. We train a GTR-T5-XL model on this expanded dataset, achieving a new benchmark of 94.7% accuracy on the SNLI dataset, 94.0% accuracy on the E-SNLI dataset, and 92.6% accuracy on the MultiNLI dataset, surpassing the previous SOTA models. This research demonstrates the potential of synthetic data augmentation in improving NLI models, offering a path forward for further advancements in natural language understanding tasks.

Results

TaskDatasetMetricValueModel
Natural Language Inferencee-SNLIAccuracy94UnitedSynT5 (3B)
Natural Language Inferencee-SNLIAccuracy89.8UnitedSynT5 (335M)
Natural Language InferenceSNLI% Test Accuracy94.7UnitedSynT5 (3B)
Natural Language InferenceSNLI% Test Accuracy93.5UnitedSynT5 (335M)
Natural Language InferenceMultiNLIMatched92.6UnitedSynT5 (3B)
Natural Language InferenceMultiNLIMatched89.8UnitedSynT5 (335M)

Related Papers

GLAD: Generalizable Tuning for Vision-Language Models2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16LRCTI: A Large Language Model-Based Framework for Multi-Step Evidence Retrieval and Reasoning in Cyber Threat Intelligence Credibility Verification2025-07-15Data Augmentation in Time Series Forecasting through Inverted Framework2025-07-15Vision Language Action Models in Robotic Manipulation: A Systematic Review2025-07-14Iceberg: Enhancing HLS Modeling with Synthetic Data2025-07-14