TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Long-Tailed Recognition via Weight Balancing

Long-Tailed Recognition via Weight Balancing

Shaden Alshammari, Yu-Xiong Wang, Deva Ramanan, Shu Kong

2022-03-27CVPR 2022 1Long-tail LearningClassification
PaperPDFCode(official)Code

Abstract

In the real open world, data tends to follow long-tailed class distributions, motivating the well-studied long-tailed recognition (LTR) problem. Naive training produces models that are biased toward common classes in terms of higher accuracy. The key to addressing LTR is to balance various aspects including data distribution, training losses, and gradients in learning. We explore an orthogonal direction, weight balancing, motivated by the empirical observation that the naively trained classifier has "artificially" larger weights in norm for common classes (because there exists abundant data to train them, unlike the rare classes). We investigate three techniques to balance weights, L2-normalization, weight decay, and MaxNorm. We first point out that L2-normalization "perfectly" balances per-class weights to be unit norm, but such a hard constraint might prevent classes from learning better classifiers. In contrast, weight decay penalizes larger weights more heavily and so learns small balanced weights; the MaxNorm constraint encourages growing small weights within a norm ball but caps all the weights by the radius. Our extensive study shows that both help learn balanced weights and greatly improve the LTR accuracy. Surprisingly, weight decay, although underexplored in LTR, significantly improves over prior work. Therefore, we adopt a two-stage training paradigm and propose a simple approach to LTR: (1) learning features using the cross-entropy loss by tuning weight decay, and (2) learning classifiers using class-balanced loss by tuning weight decay and MaxNorm. Our approach achieves the state-of-the-art accuracy on five standard benchmarks, serving as a future baseline for long-tailed recognition.

Results

TaskDatasetMetricValueModel
Image ClassificationCIFAR-100-LT (ρ=50)Error Rate42.29LTR-weight-balancing
Image ClassificationCIFAR-100-LT (ρ=10)Error Rate31.33LTR-weight-balancing
Image ClassificationImageNet-LTTop-1 Accuracy53.9LTR-weight-balancing(ResNeXt-50)
Image ClassificationCIFAR-100-LT (ρ=100)Error Rate46.45LTR-weight-balancing
Few-Shot Image ClassificationCIFAR-100-LT (ρ=50)Error Rate42.29LTR-weight-balancing
Few-Shot Image ClassificationCIFAR-100-LT (ρ=10)Error Rate31.33LTR-weight-balancing
Few-Shot Image ClassificationImageNet-LTTop-1 Accuracy53.9LTR-weight-balancing(ResNeXt-50)
Few-Shot Image ClassificationCIFAR-100-LT (ρ=100)Error Rate46.45LTR-weight-balancing
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=50)Error Rate42.29LTR-weight-balancing
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=10)Error Rate31.33LTR-weight-balancing
Generalized Few-Shot ClassificationImageNet-LTTop-1 Accuracy53.9LTR-weight-balancing(ResNeXt-50)
Generalized Few-Shot ClassificationCIFAR-100-LT (ρ=100)Error Rate46.45LTR-weight-balancing
Long-tail LearningCIFAR-100-LT (ρ=50)Error Rate42.29LTR-weight-balancing
Long-tail LearningCIFAR-100-LT (ρ=10)Error Rate31.33LTR-weight-balancing
Long-tail LearningImageNet-LTTop-1 Accuracy53.9LTR-weight-balancing(ResNeXt-50)
Long-tail LearningCIFAR-100-LT (ρ=100)Error Rate46.45LTR-weight-balancing
Generalized Few-Shot LearningCIFAR-100-LT (ρ=50)Error Rate42.29LTR-weight-balancing
Generalized Few-Shot LearningCIFAR-100-LT (ρ=10)Error Rate31.33LTR-weight-balancing
Generalized Few-Shot LearningImageNet-LTTop-1 Accuracy53.9LTR-weight-balancing(ResNeXt-50)
Generalized Few-Shot LearningCIFAR-100-LT (ρ=100)Error Rate46.45LTR-weight-balancing

Related Papers

Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Calisthenics Skills Classification through Foreground Instance Selection and Depth Estimation2025-07-16Safeguarding Federated Learning-based Road Condition Classification2025-07-16AI-Enhanced Pediatric Pneumonia Detection: A CNN-Based Approach Using Data Augmentation and Generative Adversarial Networks (GANs)2025-07-13Fuzzy Classification Aggregation for a Continuum of Agents2025-07-06Hybrid-View Attention for csPCa Classification in TRUS2025-07-04Devising a solution to the problems of Cancer awareness in Telangana2025-06-26A Semi-supervised Scalable Unified Framework for E-commerce Query Classification2025-06-26