TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/mixup: Beyond Empirical Risk Minimization

mixup: Beyond Empirical Risk Minimization

Hongyi Zhang, Moustapha Cisse, Yann N. Dauphin, David Lopez-Paz

2017-10-25ICLR 2018 1Image ClassificationDomain GeneralizationMemorizationOut-of-Distribution GeneralizationSemi-Supervised Image Classification
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode(official)CodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

Large deep neural networks are powerful, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. In this work, we propose mixup, a simple learning principle to alleviate these issues. In essence, mixup trains a neural network on convex combinations of pairs of examples and their labels. By doing so, mixup regularizes the neural network to favor simple linear behavior in-between training examples. Our experiments on the ImageNet-2012, CIFAR-10, CIFAR-100, Google commands and UCI datasets show that mixup improves the generalization of state-of-the-art neural network architectures. We also find that mixup reduces the memorization of corrupt labels, increases the robustness to adversarial examples, and stabilizes the training of generative adversarial networks.

Results

TaskDatasetMetricValueModel
Domain AdaptationImageNet-ATop-1 accuracy %6.6Mixup (ResNet-50)
Image ClassificationCIFAR-10Percentage correct97.3DenseNet-BC-190 + Mixup
Image ClassificationKuzushiji-MNISTAccuracy98.41PreActResNet-18 + Input Mixup
Image ClassificationCIFAR-100Percentage correct83.2DenseNet-BC-190 + Mixup
Image ClassificationSVHN, 250 LabelsAccuracy60.03MixUp
Image ClassificationCIFAR-10, 250 LabelsPercentage error47.43MixUp
Semi-Supervised Image ClassificationSVHN, 250 LabelsAccuracy60.03MixUp
Semi-Supervised Image ClassificationCIFAR-10, 250 LabelsPercentage error47.43MixUp
Domain GeneralizationImageNet-ATop-1 accuracy %6.6Mixup (ResNet-50)

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Simulate, Refocus and Ensemble: An Attention-Refocusing Scheme for Domain Generalization2025-07-17GLAD: Generalizable Tuning for Vision-Language Models2025-07-17MoTM: Towards a Foundation Model for Time Series Imputation based on Continuous Modeling2025-07-17