TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning From Multiple Experts: Self-paced Knowledge Disti...

Learning From Multiple Experts: Self-paced Knowledge Distillation for Long-tailed Classification

Liuyu Xiang, Guiguang Ding, Jungong Han

2020-01-06ECCV 2020 8Long-tail LearningGeneral ClassificationKnowledge Distillation
PaperPDFCode(official)

Abstract

In real-world scenarios, data tends to exhibit a long-tailed distribution, which increases the difficulty of training deep networks. In this paper, we propose a novel self-paced knowledge distillation framework, termed Learning From Multiple Experts (LFME). Our method is inspired by the observation that networks trained on less imbalanced subsets of the distribution often yield better performances than their jointly-trained counterparts. We refer to these models as 'Experts', and the proposed LFME framework aggregates the knowledge from multiple 'Experts' to learn a unified student model. Specifically, the proposed framework involves two levels of adaptive learning schedules: Self-paced Expert Selection and Curriculum Instance Selection, so that the knowledge is adaptively transferred to the 'Student'. We conduct extensive experiments and demonstrate that our method is able to achieve superior performances compared to state-of-the-art methods. We also show that our method can be easily plugged into state-of-the-art long-tailed classification algorithms for further improvements.

Results

TaskDatasetMetricValueModel
Image ClassificationPlaces-LTTop-1 Accuracy36.2LFME + OLTR
Image ClassificationImageNet-LTTop-1 Accuracy38.8LFME + OLTR
Few-Shot Image ClassificationPlaces-LTTop-1 Accuracy36.2LFME + OLTR
Few-Shot Image ClassificationImageNet-LTTop-1 Accuracy38.8LFME + OLTR
Generalized Few-Shot ClassificationPlaces-LTTop-1 Accuracy36.2LFME + OLTR
Generalized Few-Shot ClassificationImageNet-LTTop-1 Accuracy38.8LFME + OLTR
Long-tail LearningPlaces-LTTop-1 Accuracy36.2LFME + OLTR
Long-tail LearningImageNet-LTTop-1 Accuracy38.8LFME + OLTR
Generalized Few-Shot LearningPlaces-LTTop-1 Accuracy36.2LFME + OLTR
Generalized Few-Shot LearningImageNet-LTTop-1 Accuracy38.8LFME + OLTR

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training2025-07-15Feature Distillation is the Better Choice for Model-Heterogeneous Federated Learning2025-07-14KAT-V1: Kwai-AutoThink Technical Report2025-07-11Towards Collaborative Fairness in Federated Learning Under Imbalanced Covariate Shift2025-07-11SFedKD: Sequential Federated Learning with Discrepancy-Aware Multi-Teacher Knowledge Distillation2025-07-11