TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Bayesian Sparse Networks with Full Experience Rep...

Learning Bayesian Sparse Networks with Full Experience Replay for Continual Learning

Dong Gong, Qingsen Yan, Yuhang Liu, Anton Van Den Hengel, Javen Qinfeng Shi

2022-02-21CVPR 2022 1Continual LearningOpen Knowledge Graph CanonicalizationKnowledge Distillation
PaperPDF

Abstract

Continual Learning (CL) methods aim to enable machine learning models to learn new tasks without catastrophic forgetting of those that have been previously mastered. Existing CL approaches often keep a buffer of previously-seen samples, perform knowledge distillation, or use regularization techniques towards this goal. Despite their performance, they still suffer from interference across tasks which leads to catastrophic forgetting. To ameliorate this problem, we propose to only activate and select sparse neurons for learning current and past tasks at any stage. More parameters space and model capacity can thus be reserved for the future tasks. This minimizes the interference between parameters for different tasks. To do so, we propose a Sparse neural Network for Continual Learning (SNCL), which employs variational Bayesian sparsity priors on the activations of the neurons in all layers. Full Experience Replay (FER) provides effective supervision in learning the sparse activations of the neurons in different layers. A loss-aware reservoir-sampling strategy is developed to maintain the memory buffer. The proposed method is agnostic as to the network structures and the task boundaries. Experiments on different datasets show that our approach achieves state-of-the-art performance for mitigating forgetting.

Results

TaskDatasetMetricValueModel
Continual LearningTiny-ImageNet (10tasks)Average Accuracy52.85SNCL
Continual LearningTiny-ImageNet (10tasks)Average Accuracy51.78DER [buzzega2020dark]
Continual LearningTiny-ImageNet (10tasks)Average Accuracy48.64ER[riemer2018learning]
Continual LearningTiny-ImageNet (10tasks)Average Accuracy31.55iCaRL [rebuffi2017icarl]
Continual LearningTiny-ImageNet (10tasks)Average Accuracy25.33A-GEM [chaudhry2018efficient]

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17RegCL: Continual Adaptation of Segment Anything Model via Model Merging2025-07-16Information-Theoretic Generalization Bounds of Replay-based Continual Learning2025-07-16PROL : Rehearsal Free Continual Learning in Streaming Data via Prompt Online Learning2025-07-16DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Fast Last-Iterate Convergence of SGD in the Smooth Interpolation Regime2025-07-15A Neural Network Model of Complementary Learning Systems: Pattern Separation and Completion for Continual Learning2025-07-15