TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Class-Balanced Distillation for Long-Tailed Visual Recogni...

Class-Balanced Distillation for Long-Tailed Visual Recognition

Ahmet Iscen, André Araujo, Boqing Gong, Cordelia Schmid

2021-04-12Image ClassificationLong-tail LearningKnowledge Distillation
PaperPDFCodeCode(official)Code(official)

Abstract

Real-world imagery is often characterized by a significant imbalance of the number of images per class, leading to long-tailed distributions. An effective and simple approach to long-tailed visual recognition is to learn feature representations and a classifier separately, with instance and class-balanced sampling, respectively. In this work, we introduce a new framework, by making the key observation that a feature representation learned with instance sampling is far from optimal in a long-tailed setting. Our main contribution is a new training method, referred to as Class-Balanced Distillation (CBD), that leverages knowledge distillation to enhance feature representations. CBD allows the feature representation to evolve in the second training stage, guided by the teacher learned in the first stage. The second stage uses class-balanced sampling, in order to focus on under-represented classes. This framework can naturally accommodate the usage of multiple teachers, unlocking the information from an ensemble of models to enhance recognition capabilities. Our experiments show that the proposed technique consistently outperforms the state of the art on long-tailed recognition benchmarks such as ImageNet-LT, iNaturalist17 and iNaturalist18.

Results

TaskDatasetMetricValueModel
Image ClassificationImageNet-LTTop-1 Accuracy57.7CBD-ENS (ResNet-152)
Image ClassificationImageNet-LTTop-1 Accuracy55.6CBD-ENS (ResNet-50)
Few-Shot Image ClassificationImageNet-LTTop-1 Accuracy57.7CBD-ENS (ResNet-152)
Few-Shot Image ClassificationImageNet-LTTop-1 Accuracy55.6CBD-ENS (ResNet-50)
Generalized Few-Shot ClassificationImageNet-LTTop-1 Accuracy57.7CBD-ENS (ResNet-152)
Generalized Few-Shot ClassificationImageNet-LTTop-1 Accuracy55.6CBD-ENS (ResNet-50)
Long-tail LearningImageNet-LTTop-1 Accuracy57.7CBD-ENS (ResNet-152)
Long-tail LearningImageNet-LTTop-1 Accuracy55.6CBD-ENS (ResNet-50)
Generalized Few-Shot LearningImageNet-LTTop-1 Accuracy57.7CBD-ENS (ResNet-152)
Generalized Few-Shot LearningImageNet-LTTop-1 Accuracy55.6CBD-ENS (ResNet-50)

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16