TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Generalized Cross Entropy Loss for Training Deep Neural Ne...

Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels

Zhilu Zhang, Mert R. Sabuncu

2018-05-20NeurIPS 2018 12Image ClassificationLearning with noisy labels
PaperPDFCodeCodeCodeCode

Abstract

Deep neural networks (DNNs) have achieved tremendous success in a variety of applications across many disciplines. Yet, their superior performance comes with the expensive cost of requiring correctly annotated large-scale datasets. Moreover, due to DNNs' rich capacity, errors in training labels can hamper performance. To combat this problem, mean absolute error (MAE) has recently been proposed as a noise-robust alternative to the commonly-used categorical cross entropy (CCE) loss. However, as we show in this paper, MAE can perform poorly with DNNs and challenging datasets. Here, we present a theoretically grounded set of noise-robust loss functions that can be seen as a generalization of MAE and CCE. Proposed loss functions can be readily applied with any existing DNN architecture and algorithm, while yielding good performance in a wide range of noisy label scenarios. We report results from experiments conducted with CIFAR-10, CIFAR-100 and FASHION-MNIST datasets and synthetically generated noisy labels.

Results

TaskDatasetMetricValueModel
Image ClassificationCIFAR-10N-Random2Accuracy (mean)87.7GCE
Image ClassificationCIFAR-10N-Random3Accuracy (mean)87.58GCE
Image ClassificationCIFAR-10N-AggregateAccuracy (mean)87.85GCE
Image ClassificationCIFAR-10N-Random1Accuracy (mean)87.61GCE
Image ClassificationCIFAR-100NAccuracy (mean)56.73GCE
Image ClassificationCIFAR-10N-WorstAccuracy (mean)80.66GCE
Document Text ClassificationCIFAR-10N-Random2Accuracy (mean)87.7GCE
Document Text ClassificationCIFAR-10N-Random3Accuracy (mean)87.58GCE
Document Text ClassificationCIFAR-10N-AggregateAccuracy (mean)87.85GCE
Document Text ClassificationCIFAR-10N-Random1Accuracy (mean)87.61GCE
Document Text ClassificationCIFAR-100NAccuracy (mean)56.73GCE
Document Text ClassificationCIFAR-10N-WorstAccuracy (mean)80.66GCE

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17CLID-MU: Cross-Layer Information Divergence Based Meta Update Strategy for Learning with Noisy Labels2025-07-16Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Transferring Styles for Reduced Texture Bias and Improved Robustness in Semantic Segmentation Networks2025-07-14