TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Sparse Training via Boosting Pruning Plasticity with Neuro...

Sparse Training via Boosting Pruning Plasticity with Neuroregeneration

Shiwei Liu, Tianlong Chen, Xiaohan Chen, Zahra Atashgahi, Lu Yin, Huanyu Kou, Li Shen, Mykola Pechenizkiy, Zhangyang Wang, Decebal Constantin Mocanu

2021-06-19NeurIPS 2021 12Network PruningSparse Learning
PaperPDFCodeCode(official)

Abstract

Works on lottery ticket hypothesis (LTH) and single-shot network pruning (SNIP) have raised a lot of attention currently on post-training pruning (iterative magnitude pruning), and before-training pruning (pruning at initialization). The former method suffers from an extremely large computation cost and the latter usually struggles with insufficient performance. In comparison, during-training pruning, a class of pruning methods that simultaneously enjoys the training/inference efficiency and the comparable performance, temporarily, has been less explored. To better understand during-training pruning, we quantitatively study the effect of pruning throughout training from the perspective of pruning plasticity (the ability of the pruned networks to recover the original performance). Pruning plasticity can help explain several other empirical observations about neural network pruning in literature. We further find that pruning plasticity can be substantially improved by injecting a brain-inspired mechanism called neuroregeneration, i.e., to regenerate the same number of connections as pruned. We design a novel gradual magnitude pruning (GMP) method, named gradual pruning with zero-cost neuroregeneration (\textbf{GraNet}), that advances state of the art. Perhaps most impressively, its sparse-to-sparse version for the first time boosts the sparse-to-sparse training performance over various dense-to-sparse methods with ResNet-50 on ImageNet without extending the training time. We release all codes in https://github.com/Shiweiliuiiiiiii/GraNet.

Results

TaskDatasetMetricValueModel
Sparse LearningImageNetTop-1 Accuracy76Resnet-50: 80% Sparse 100 epochs
Sparse LearningImageNetTop-1 Accuracy74.5Resnet-50: 90% Sparse 100 epochs

Related Papers

Scalable Subset Selection in Linear Mixed Models2025-06-25Hyperpruning: Efficient Search through Pruned Variants of Recurrent Neural Networks Leveraging Lyapunov Spectrum2025-06-09Learning Sparsity for Effective and Efficient Music Performance Question Answering2025-06-02TSENOR: Highly-Efficient Algorithm for Finding Transposable N:M Sparse Masks2025-05-29Hierarchical Safety Realignment: Lightweight Restoration of Safety in Pruned Large Vision-Language Models2025-05-22Adaptive Pruning of Deep Neural Networks for Resource-Aware Embedded Intrusion Detection on the Edge2025-05-20Bi-LSTM based Multi-Agent DRL with Computation-aware Pruning for Agent Twins Migration in Vehicular Embodied AI Networks2025-05-09Guiding Evolutionary AutoEncoder Training with Activation-Based Pruning Operators2025-05-08