TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning with Instance-Dependent Label Noise: A Sample Sie...

Learning with Instance-Dependent Label Noise: A Sample Sieve Approach

Hao Cheng, Zhaowei Zhu, Xingyu Li, Yifei Gong, Xing Sun, Yang Liu

2020-10-05ICLR 2021 1Image ClassificationImage Classification with Label NoiseLearning with noisy labels
PaperPDFCode(official)

Abstract

Human-annotated labels are often prone to noise, and the presence of such noise will degrade the performance of the resulting deep neural network (DNN) models. Much of the literature (with several recent exceptions) of learning with noisy labels focuses on the case when the label noise is independent of features. Practically, annotations errors tend to be instance-dependent and often depend on the difficulty levels of recognizing a certain task. Applying existing results from instance-independent settings would require a significant amount of estimation of noise rates. Therefore, providing theoretically rigorous solutions for learning with instance-dependent label noise remains a challenge. In this paper, we propose CORES$^{2}$ (COnfidence REgularized Sample Sieve), which progressively sieves out corrupted examples. The implementation of CORES$^{2}$ does not require specifying noise rates and yet we are able to provide theoretical guarantees of CORES$^{2}$ in filtering out the corrupted examples. This high-quality sample sieve allows us to treat clean examples and the corrupted ones separately in training a DNN solution, and such a separation is shown to be advantageous in the instance-dependent noise setting. We demonstrate the performance of CORES$^{2}$ on CIFAR10 and CIFAR100 datasets with synthetic instance-dependent label noise and Clothing1M with real-world human noise. As of independent interests, our sample sieve provides a generic machinery for anatomizing noisy datasets and provides a flexible interface for various robust training techniques to further improve the performance. Code is available at https://github.com/UCSC-REAL/cores.

Results

TaskDatasetMetricValueModel
Image ClassificationCIFAR-10N-Random2Accuracy (mean)94.88CORES*
Image ClassificationCIFAR-10N-Random2Accuracy (mean)89.91CORES
Image ClassificationCIFAR-10N-Random3Accuracy (mean)94.74CORES*
Image ClassificationCIFAR-10N-Random3Accuracy (mean)89.79CORES
Image ClassificationCIFAR-10N-AggregateAccuracy (mean)95.25CORES*
Image ClassificationCIFAR-10N-AggregateAccuracy (mean)91.23CORES
Image ClassificationCIFAR-10N-Random1Accuracy (mean)94.45CORES*
Image ClassificationCIFAR-10N-Random1Accuracy (mean)89.66CORES
Image ClassificationCIFAR-100NAccuracy (mean)61.15CORES
Image ClassificationCIFAR-100NAccuracy (mean)55.72CORES*
Image ClassificationCIFAR-10N-WorstAccuracy (mean)91.66CORES*
Image ClassificationCIFAR-10N-WorstAccuracy (mean)83.6CORES
Document Text ClassificationCIFAR-10N-Random2Accuracy (mean)94.88CORES*
Document Text ClassificationCIFAR-10N-Random2Accuracy (mean)89.91CORES
Document Text ClassificationCIFAR-10N-Random3Accuracy (mean)94.74CORES*
Document Text ClassificationCIFAR-10N-Random3Accuracy (mean)89.79CORES
Document Text ClassificationCIFAR-10N-AggregateAccuracy (mean)95.25CORES*
Document Text ClassificationCIFAR-10N-AggregateAccuracy (mean)91.23CORES
Document Text ClassificationCIFAR-10N-Random1Accuracy (mean)94.45CORES*
Document Text ClassificationCIFAR-10N-Random1Accuracy (mean)89.66CORES
Document Text ClassificationCIFAR-100NAccuracy (mean)61.15CORES
Document Text ClassificationCIFAR-100NAccuracy (mean)55.72CORES*
Document Text ClassificationCIFAR-10N-WorstAccuracy (mean)91.66CORES*
Document Text ClassificationCIFAR-10N-WorstAccuracy (mean)83.6CORES

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17CLID-MU: Cross-Layer Information Divergence Based Meta Update Strategy for Learning with Noisy Labels2025-07-16Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Transferring Styles for Reduced Texture Bias and Improved Robustness in Semantic Segmentation Networks2025-07-14