TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Pushing the Limits of Simple Pipelines for Few-Shot Learni...

Pushing the Limits of Simple Pipelines for Few-Shot Learning: External Data and Fine-Tuning Make a Difference

Shell Xu Hu, Da Li, Jan Stühmer, Minyoung Kim, Timothy M. Hospedales

2022-04-15CVPR 2022 1Few-Shot LearningMeta-LearningImage ClassificationTransfer LearningFew-Shot Image Classification
PaperPDFCode(official)

Abstract

Few-shot learning (FSL) is an important and topical problem in computer vision that has motivated extensive research into numerous methods spanning from sophisticated meta-learning methods to simple transfer learning baselines. We seek to push the limits of a simple-but-effective pipeline for more realistic and practical settings of few-shot image classification. To this end, we explore few-shot learning from the perspective of neural network architecture, as well as a three stage pipeline of network updates under different data supplies, where unsupervised external data is considered for pre-training, base categories are used to simulate few-shot tasks for meta-training, and the scarcely labelled data of an novel task is taken for fine-tuning. We investigate questions such as: (1) How pre-training on external data benefits FSL? (2) How state-of-the-art transformer architectures can be exploited? and (3) How fine-tuning mitigates domain shift? Ultimately, we show that a simple transformer-based pipeline yields surprisingly good performance on standard benchmarks such as Mini-ImageNet, CIFAR-FS, CDFSL and Meta-Dataset. Our code and demo are available at https://hushell.github.io/pmf.

Results

TaskDatasetMetricValueModel
Image ClassificationMeta-DatasetAccuracy84.75P>M>F (P=DINO-ViT-base, M=ProtoNet)
Image ClassificationCIFAR-FS 5-way (1-shot)Accuracy84.3P>M>F (P=DINO-ViT-base, M=ProtoNet)
Image ClassificationMini-Imagenet 5-way (5-shot)Accuracy98.4P>M>F (P=DINO-ViT-base, M=ProtoNet)
Image ClassificationMini-Imagenet 5-way (1-shot)Accuracy95.3P>M>F (P=DINO-ViT-base, M=ProtoNet)
Image ClassificationCIFAR-FS 5-way (5-shot)Accuracy92.2P>M>F (P=DINO-ViT-base, M=ProtoNet)
Few-Shot Image ClassificationMeta-DatasetAccuracy84.75P>M>F (P=DINO-ViT-base, M=ProtoNet)
Few-Shot Image ClassificationCIFAR-FS 5-way (1-shot)Accuracy84.3P>M>F (P=DINO-ViT-base, M=ProtoNet)
Few-Shot Image ClassificationMini-Imagenet 5-way (5-shot)Accuracy98.4P>M>F (P=DINO-ViT-base, M=ProtoNet)
Few-Shot Image ClassificationMini-Imagenet 5-way (1-shot)Accuracy95.3P>M>F (P=DINO-ViT-base, M=ProtoNet)
Few-Shot Image ClassificationCIFAR-FS 5-way (5-shot)Accuracy92.2P>M>F (P=DINO-ViT-base, M=ProtoNet)

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18GLAD: Generalizable Tuning for Vision-Language Models2025-07-17Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17