TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Disentangling Label Distribution for Long-tailed Visual Re...

Disentangling Label Distribution for Long-tailed Visual Recognition

Youngkyu Hong, Seungju Han, Kwanghee Choi, Seokjun Seo, Beomsu Kim, Buru Chang

2020-12-01CVPR 2021 1Image ClassificationLong-tail LearningPrediction
PaperPDFCodeCode(official)

Abstract

The current evaluation protocol of long-tailed visual recognition trains the classification model on the long-tailed source label distribution and evaluates its performance on the uniform target label distribution. Such protocol has questionable practicality since the target may also be long-tailed. Therefore, we formulate long-tailed visual recognition as a label shift problem where the target and source label distributions are different. One of the significant hurdles in dealing with the label shift problem is the entanglement between the source label distribution and the model prediction. In this paper, we focus on disentangling the source label distribution from the model prediction. We first introduce a simple but overlooked baseline method that matches the target label distribution by post-processing the model prediction trained by the cross-entropy loss and the Softmax function. Although this method surpasses state-of-the-art methods on benchmark datasets, it can be further improved by directly disentangling the source label distribution from the model prediction in the training phase. Thus, we propose a novel method, LAbel distribution DisEntangling (LADE) loss based on the optimal bound of Donsker-Varadhan representation. LADE achieves state-of-the-art performance on benchmark datasets such as CIFAR-100-LT, Places-LT, ImageNet-LT, and iNaturalist 2018. Moreover, LADE outperforms existing methods on various shifted target label distributions, showing the general adaptability of our proposed method.

Results

TaskDatasetMetricValueModel
Image ClassificationPlaces-LTTop-1 Accuracy38.8LADE
Image ClassificationCIFAR-10-LT (ρ=10)Error Rate11.22LADE
Image ClassificationCIFAR-100-LT (ρ=10)Error Rate38.3LADE
Image ClassificationImageNet-LTTop-1 Accuracy53LADE
Image ClassificationCIFAR-100-LT (ρ=100)Error Rate54.6LADE
Few-Shot Image ClassificationPlaces-LTTop-1 Accuracy38.8LADE
Few-Shot Image ClassificationCIFAR-10-LT (ρ=10)Error Rate11.22LADE
Few-Shot Image ClassificationCIFAR-100-LT (ρ=10)Error Rate38.3LADE
Few-Shot Image ClassificationImageNet-LTTop-1 Accuracy53LADE
Few-Shot Image ClassificationCIFAR-100-LT (ρ=100)Error Rate54.6LADE
Generalized Few-Shot ClassificationPlaces-LTTop-1 Accuracy38.8LADE
Generalized Few-Shot ClassificationCIFAR-10-LT (ρ=10)Error Rate11.22LADE
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=10)Error Rate38.3LADE
Generalized Few-Shot ClassificationImageNet-LTTop-1 Accuracy53LADE
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=100)Error Rate54.6LADE
Long-tail LearningPlaces-LTTop-1 Accuracy38.8LADE
Long-tail LearningCIFAR-10-LT (ρ=10)Error Rate11.22LADE
Long-tail LearningCIFAR-100-LT (ρ=10)Error Rate38.3LADE
Long-tail LearningImageNet-LTTop-1 Accuracy53LADE
Long-tail LearningCIFAR-100-LT (ρ=100)Error Rate54.6LADE
Generalized Few-Shot LearningPlaces-LTTop-1 Accuracy38.8LADE
Generalized Few-Shot LearningCIFAR-10-LT (ρ=10)Error Rate11.22LADE
Generalized Few-Shot LearningCIFAR-100-LT (ρ=10)Error Rate38.3LADE
Generalized Few-Shot LearningImageNet-LTTop-1 Accuracy53LADE
Generalized Few-Shot LearningCIFAR-100-LT (ρ=100)Error Rate54.6LADE

Related Papers

Multi-Strategy Improved Snake Optimizer Accelerated CNN-LSTM-Attention-Adaboost for Trajectory Prediction2025-07-21Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Hashed Watermark as a Filter: Defeating Forging and Overwriting Attacks in Weight-based Neural Network Watermarking2025-07-15Generative Click-through Rate Prediction with Applications to Search Advertising2025-07-15