TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Self-supervised Knowledge Distillation for Few-shot Learning

Self-supervised Knowledge Distillation for Few-shot Learning

Jathushan Rajasegaran, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan, Mubarak Shah

2020-06-17Few-Shot LearningMeta-LearningMetric LearningFew-Shot Image ClassificationKnowledge Distillation
PaperPDFCodeCode(official)

Abstract

Real-world contains an overwhelmingly large number of object classes, learning all of which at once is infeasible. Few shot learning is a promising learning paradigm due to its ability to learn out of order distributions quickly with only a few samples. Recent works [7, 41] show that simply learning a good feature embedding can outperform more sophisticated meta-learning and metric learning algorithms for few-shot learning. In this paper, we propose a simple approach to improve the representation capacity of deep neural networks for few-shot learning tasks. We follow a two-stage learning process: First, we train a neural network to maximize the entropy of the feature embedding, thus creating an optimal output manifold using a self-supervised auxiliary loss. In the second stage, we minimize the entropy on feature embedding by bringing self-supervised twins together, while constraining the manifold with student-teacher distillation. Our experiments show that, even in the first stage, self-supervision can outperform current state-of-the-art methods, with further gains achieved by our second stage distillation process. Our codes are available at: https://github.com/brjathu/SKD.

Results

TaskDatasetMetricValueModel
Image ClassificationCIFAR-FS 5-way (1-shot)Accuracy76.9SKD
Image ClassificationMini-Imagenet 5-way (5-shot)Accuracy83.54SKD
Image ClassificationMini-Imagenet 5-way (1-shot)Accuracy67.04SKD
Image ClassificationFC100 5-way (5-shot)Accuracy63.1SKD
Image ClassificationFC100 5-way (1-shot)Accuracy46.5SKD
Image ClassificationTiered ImageNet 5-way (1-shot)Accuracy72.03SKD
Image ClassificationTiered ImageNet 5-way (5-shot)Accuracy86.66SKD
Image ClassificationCIFAR-FS 5-way (5-shot)Accuracy88.9SKD
Few-Shot Image ClassificationCIFAR-FS 5-way (1-shot)Accuracy76.9SKD
Few-Shot Image ClassificationMini-Imagenet 5-way (5-shot)Accuracy83.54SKD
Few-Shot Image ClassificationMini-Imagenet 5-way (1-shot)Accuracy67.04SKD
Few-Shot Image ClassificationFC100 5-way (5-shot)Accuracy63.1SKD
Few-Shot Image ClassificationFC100 5-way (1-shot)Accuracy46.5SKD
Few-Shot Image ClassificationTiered ImageNet 5-way (1-shot)Accuracy72.03SKD
Few-Shot Image ClassificationTiered ImageNet 5-way (5-shot)Accuracy86.66SKD
Few-Shot Image ClassificationCIFAR-FS 5-way (5-shot)Accuracy88.9SKD

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21GLAD: Generalizable Tuning for Vision-Language Models2025-07-17Unsupervised Ground Metric Learning2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16CLID-MU: Cross-Layer Information Divergence Based Meta Update Strategy for Learning with Noisy Labels2025-07-16DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16