TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Unsupervised Data Augmentation for Consistency Training

Unsupervised Data Augmentation for Consistency Training

Qizhe Xie, Zihang Dai, Eduard Hovy, Minh-Thang Luong, Quoc V. Le

2019-04-29NeurIPS 2020 12Text ClassificationImage AugmentationImage ClassificationSentiment AnalysisData AugmentationTransfer LearningSemi-Supervised Image Classification
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCode

Abstract

Semi-supervised learning lately has shown much promise in improving deep learning models when labeled data is scarce. Common among recent approaches is the use of consistency training on a large amount of unlabeled data to constrain model predictions to be invariant to input noise. In this work, we present a new perspective on how to effectively noise unlabeled examples and argue that the quality of noising, specifically those produced by advanced data augmentation methods, plays a crucial role in semi-supervised learning. By substituting simple noising operations with advanced data augmentation methods such as RandAugment and back-translation, our method brings substantial improvements across six language and three vision tasks under the same consistency training framework. On the IMDb text classification dataset, with only 20 labeled examples, our method achieves an error rate of 4.20, outperforming the state-of-the-art model trained on 25,000 labeled examples. On a standard semi-supervised learning benchmark, CIFAR-10, our method outperforms all previous approaches and achieves an error rate of 5.43 with only 250 examples. Our method also combines well with transfer learning, e.g., when finetuning from BERT, and yields improvements in high-data regime, such as ImageNet, whether when there is only 10% labeled data or when a full labeled set with 1.3M extra unlabeled examples is used. Code is available at https://github.com/google-research/uda.

Results

TaskDatasetMetricValueModel
Sentiment AnalysisAmazon Review PolarityAccuracy97.37BERT large
Sentiment AnalysisAmazon Review PolarityAccuracy96.5BERT large finetune UDA
Sentiment AnalysisYelp Fine-grained classificationError29.32BERT large
Sentiment AnalysisYelp Fine-grained classificationError32.08BERT large finetune UDA
Sentiment AnalysisYelp Binary classificationError1.89BERT large
Sentiment AnalysisYelp Binary classificationError2.05BERT large finetune UDA
Sentiment AnalysisIMDbAccuracy95.8BERT large finetune UDA
Sentiment AnalysisIMDbAccuracy95.49BERT large
Sentiment AnalysisAmazon Review FullAccuracy65.83BERT large
Sentiment AnalysisAmazon Review FullAccuracy62.88BERT large finetune UDA
Text ClassificationDBpediaError0.68BERT large
Text ClassificationDBpediaError1.09BERT large UDA
Text ClassificationAmazon-5Error37.12BERT Finetune + UDA
Text ClassificationAmazon-2Error3.5BERT Finetune + UDA
Image ClassificationCIFAR-10, 4000 LabelsPercentage error5.27UDA
Image ClassificationImageNet - 10% labeled dataTop 5 Accuracy88.52UDA
Image ClassificationSVHN, 1000 labelsAccuracy97.54UDA
ClassificationDBpediaError0.68BERT large
ClassificationDBpediaError1.09BERT large UDA
ClassificationAmazon-5Error37.12BERT Finetune + UDA
ClassificationAmazon-2Error3.5BERT Finetune + UDA
Semi-Supervised Image ClassificationCIFAR-10, 4000 LabelsPercentage error5.27UDA
Semi-Supervised Image ClassificationImageNet - 10% labeled dataTop 5 Accuracy88.52UDA
Semi-Supervised Image ClassificationSVHN, 1000 labelsAccuracy97.54UDA

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Making Language Model a Hierarchical Classifier and Generator2025-07-17Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17AdaptiSent: Context-Aware Adaptive Attention for Multimodal Aspect-Based Sentiment Analysis2025-07-17