TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Unified Generalization Analysis of Re-Weighting and Logi...

A Unified Generalization Analysis of Re-Weighting and Logit-Adjustment for Imbalanced Learning

Zitai Wang, Qianqian Xu, Zhiyong Yang, Yuan He, Xiaochun Cao, Qingming Huang

2023-10-07NeurIPS 2023 11
PaperPDFCode

Abstract

Real-world datasets are typically imbalanced in the sense that only a few classes have numerous samples, while many classes are associated with only a few samples. As a result, a na\"ive ERM learning process will be biased towards the majority classes, making it difficult to generalize to the minority classes. To address this issue, one simple but effective approach is to modify the loss function to emphasize the learning on minority classes, such as re-weighting the losses or adjusting the logits via class-dependent terms. However, existing generalization analysis of such losses is still coarse-grained and fragmented, failing to explain some empirical results. To bridge this gap, we propose a novel technique named data-dependent contraction to capture how these modified losses handle different classes. On top of this technique, a fine-grained generalization bound is established for imbalanced learning, which helps reveal the mystery of re-weighting and logit-adjustment in a unified manner. Furthermore, a principled learning algorithm is developed based on the theoretical insights. Finally, the empirical results on benchmark datasets not only validate the theoretical results but also demonstrate the effectiveness of the proposed method.

Results

TaskDatasetMetricValueModel
Image ClassificationCIFAR-10-LT (ρ=10)Error Rate8.18VS + ADRW + TLA
Image ClassificationCIFAR-100-LT (ρ=10)Error Rate34.41VS + ADRW + TLA
Image ClassificationCIFAR-100-LT (ρ=100)Error Rate46.95VS + ADRW + TLA
Image ClassificationCIFAR-10-LT (ρ=100)Error Rate13.58VS + ADRW + TLA
Few-Shot Image ClassificationCIFAR-10-LT (ρ=10)Error Rate8.18VS + ADRW + TLA
Few-Shot Image ClassificationCIFAR-100-LT (ρ=10)Error Rate34.41VS + ADRW + TLA
Few-Shot Image ClassificationCIFAR-100-LT (ρ=100)Error Rate46.95VS + ADRW + TLA
Few-Shot Image ClassificationCIFAR-10-LT (ρ=100)Error Rate13.58VS + ADRW + TLA
Generalized Few-Shot ClassificationCIFAR-10-LT (ρ=10)Error Rate8.18VS + ADRW + TLA
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=10)Error Rate34.41VS + ADRW + TLA
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=100)Error Rate46.95VS + ADRW + TLA
Generalized Few-Shot ClassificationCIFAR-10-LT (ρ=100)Error Rate13.58VS + ADRW + TLA
Long-tail LearningCIFAR-10-LT (ρ=10)Error Rate8.18VS + ADRW + TLA
Long-tail LearningCIFAR-100-LT (ρ=10)Error Rate34.41VS + ADRW + TLA
Long-tail LearningCIFAR-100-LT (ρ=100)Error Rate46.95VS + ADRW + TLA
Long-tail LearningCIFAR-10-LT (ρ=100)Error Rate13.58VS + ADRW + TLA
Generalized Few-Shot LearningCIFAR-10-LT (ρ=10)Error Rate8.18VS + ADRW + TLA
Generalized Few-Shot LearningCIFAR-100-LT (ρ=10)Error Rate34.41VS + ADRW + TLA
Generalized Few-Shot LearningCIFAR-100-LT (ρ=100)Error Rate46.95VS + ADRW + TLA
Generalized Few-Shot LearningCIFAR-10-LT (ρ=100)Error Rate13.58VS + ADRW + TLA