TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Memory-based Jitter: Improving Visual Recognition on Long-...

Memory-based Jitter: Improving Visual Recognition on Long-tailed Data with Diversity In Memory

Jialun Liu, Jingwei Zhang, Yi Yang, Wenhui Li, Chi Zhang, Yifan Sun

2020-08-22Image ClassificationLong-tail LearningMetric LearningData AugmentationPerson Re-IdentificationGeneral Classification
PaperPDF

Abstract

This paper considers deep visual recognition on long-tailed data. To be general, we consider two applied scenarios, \ie, deep classification and deep metric learning. Under the long-tailed data distribution, the majority classes (\ie, tail classes) only occupy relatively few samples and are prone to lack of within-class diversity. A radical solution is to augment the tail classes with higher diversity. To this end, we introduce a simple and reliable method named Memory-based Jitter (MBJ). We observe that during training, the deep model constantly changes its parameters after every iteration, yielding the phenomenon of \emph{weight jitters}. Consequentially, given a same image as the input, two historical editions of the model generate two different features in the deeply-embedded space, resulting in \emph{feature jitters}. Using a memory bank, we collect these (model or feature) jitters across multiple training iterations and get the so-called Memory-based Jitter. The accumulated jitters enhance the within-class diversity for the tail classes and consequentially improves long-tailed visual recognition. With slight modifications, MBJ is applicable for two fundamental visual recognition tasks, \emph{i.e.}, deep image classification and deep metric learning (on long-tailed data). Extensive experiments on five long-tailed classification benchmarks and two deep metric learning benchmarks demonstrate significant improvement. Moreover, the achieved performance are on par with the state of the art on both tasks.

Results

TaskDatasetMetricValueModel
Image ClassificationCIFAR-100-LT (ρ=100)Error Rate54.2MBJ
Few-Shot Image ClassificationCIFAR-100-LT (ρ=100)Error Rate54.2MBJ
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=100)Error Rate54.2MBJ
Long-tail LearningCIFAR-100-LT (ρ=100)Error Rate54.2MBJ
Generalized Few-Shot LearningCIFAR-100-LT (ρ=100)Error Rate54.2MBJ

Related Papers

Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17Unsupervised Ground Metric Learning2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17