TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/RandAugment: Practical automated data augmentation with a ...

RandAugment: Practical automated data augmentation with a reduced search space

Ekin D. Cubuk, Barret Zoph, Jonathon Shlens, Quoc V. Le

2019-09-30NeurIPS 2020 12Image ClassificationData AugmentationDomain Generalizationobject-detectionObject Detection
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Recent work has shown that data augmentation has the potential to significantly improve the generalization of deep learning models. Recently, automated augmentation strategies have led to state-of-the-art results in image classification and object detection. While these strategies were optimized for improving validation accuracy, they also led to state-of-the-art results in semi-supervised learning and improved robustness to common corruptions of images. An obstacle to a large-scale adoption of these methods is a separate search phase which increases the training complexity and may substantially increase the computational cost. Additionally, due to the separate search phase, these approaches are unable to adjust the regularization strength based on model or dataset size. Automated augmentation policies are often found by training small models on small datasets and subsequently applied to train larger models. In this work, we remove both of these obstacles. RandAugment has a significantly reduced search space which allows it to be trained on the target task with no need for a separate proxy task. Furthermore, due to the parameterization, the regularization strength may be tailored to different model and dataset sizes. RandAugment can be used uniformly across different tasks and datasets and works out of the box, matching or surpassing all previous automated augmentation approaches on CIFAR-10/100, SVHN, and ImageNet. On the ImageNet dataset we achieve 85.0% accuracy, a 0.6% increase over the previous state-of-the-art and 1.0% increase over baseline augmentation. On object detection, RandAugment leads to 1.0-1.3% improvement over baseline augmentation, and is within 0.3% mAP of AutoAugment on COCO. Finally, due to its interpretable hyperparameter, RandAugment may be used to investigate the role of data augmentation with varying model and dataset size. Code is available online.

Results

TaskDatasetMetricValueModel
Domain AdaptationVizWiz-ClassificationAccuracy - All Images45EfficientNet-B7 (randaug)
Domain AdaptationVizWiz-ClassificationAccuracy - Clean Images48.7EfficientNet-B7 (randaug)
Domain AdaptationVizWiz-ClassificationAccuracy - Corrupted Images38.9EfficientNet-B7 (randaug)
Domain AdaptationVizWiz-ClassificationAccuracy - All Images42.1EfficientNet-B5 (randaug)
Domain AdaptationVizWiz-ClassificationAccuracy - Clean Images47.3EfficientNet-B5 (randaug)
Domain AdaptationVizWiz-ClassificationAccuracy - Corrupted Images35.5EfficientNet-B5 (randaug)
Data AugmentationImageNetAccuracy (%)77.6ResNet-50 (RA)
Domain GeneralizationVizWiz-ClassificationAccuracy - All Images45EfficientNet-B7 (randaug)
Domain GeneralizationVizWiz-ClassificationAccuracy - Clean Images48.7EfficientNet-B7 (randaug)
Domain GeneralizationVizWiz-ClassificationAccuracy - Corrupted Images38.9EfficientNet-B7 (randaug)
Domain GeneralizationVizWiz-ClassificationAccuracy - All Images42.1EfficientNet-B5 (randaug)
Domain GeneralizationVizWiz-ClassificationAccuracy - Clean Images47.3EfficientNet-B5 (randaug)
Domain GeneralizationVizWiz-ClassificationAccuracy - Corrupted Images35.5EfficientNet-B5 (randaug)

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17