TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/How to train your MAML

How to train your MAML

Antreas Antoniou, Harrison Edwards, Amos Storkey

2018-10-22ICLR 2019 5Few-Shot LearningMeta-LearningFew-Shot Image Classification
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

The field of few-shot learning has recently seen substantial advancements. Most of these advancements came from casting few-shot learning as a meta-learning problem. Model Agnostic Meta Learning or MAML is currently one of the best approaches for few-shot learning via meta-learning. MAML is simple, elegant and very powerful, however, it has a variety of issues, such as being very sensitive to neural network architectures, often leading to instability during training, requiring arduous hyperparameter searches to stabilize training and achieve high generalization and being very computationally expensive at both training and inference times. In this paper, we propose various modifications to MAML that not only stabilize the system, but also substantially improve the generalization performance, convergence speed and computational overhead of MAML, which we call MAML++.

Results

TaskDatasetMetricValueModel
Image ClassificationOMNIGLOT - 1-Shot, 5-wayAccuracy99.47MAML++
Image ClassificationMini-Imagenet 5-way (5-shot)Accuracy67.15MAML++
Image ClassificationMini-Imagenet 5-way (1-shot)Accuracy52.4MAML++
Image ClassificationOMNIGLOT - 1-Shot, 20-wayAccuracy97.65MAML++
Few-Shot Image ClassificationOMNIGLOT - 1-Shot, 5-wayAccuracy99.47MAML++
Few-Shot Image ClassificationMini-Imagenet 5-way (5-shot)Accuracy67.15MAML++
Few-Shot Image ClassificationMini-Imagenet 5-way (1-shot)Accuracy52.4MAML++
Few-Shot Image ClassificationOMNIGLOT - 1-Shot, 20-wayAccuracy97.65MAML++

Related Papers

GLAD: Generalizable Tuning for Vision-Language Models2025-07-17Are encoders able to learn landmarkers for warm-starting of Hyperparameter Optimization?2025-07-16Imbalanced Regression Pipeline Recommendation2025-07-16CLID-MU: Cross-Layer Information Divergence Based Meta Update Strategy for Learning with Noisy Labels2025-07-16Mixture of Experts in Large Language Models2025-07-15Iceberg: Enhancing HLS Modeling with Synthetic Data2025-07-14Meta-Reinforcement Learning for Fast and Data-Efficient Spectrum Allocation in Dynamic Wireless Networks2025-07-13ViT-ProtoNet for Few-Shot Image Classification: A Multi-Benchmark Evaluation2025-07-12