TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Divide and not forget: Ensemble of selectively trained exp...

Divide and not forget: Ensemble of selectively trained experts in Continual Learning

Grzegorz Rypeść, Sebastian Cygert, Valeriya Khan, Tomasz Trzciński, Bartosz Zieliński, Bartłomiej Twardowski

2024-01-18Continual LearningClass Incremental Learningclass-incremental learningIncremental Learning
PaperPDFCode(official)

Abstract

Class-incremental learning is becoming more popular as it helps models widen their applicability while not forgetting what they already know. A trend in this area is to use a mixture-of-expert technique, where different models work together to solve the task. However, the experts are usually trained all at once using whole task data, which makes them all prone to forgetting and increasing computational burden. To address this limitation, we introduce a novel approach named SEED. SEED selects only one, the most optimal expert for a considered task, and uses data from this task to fine-tune only this expert. For this purpose, each expert represents each class with a Gaussian distribution, and the optimal expert is selected based on the similarity of those distributions. Consequently, SEED increases diversity and heterogeneity within the experts while maintaining the high stability of this ensemble method. The extensive experiments demonstrate that SEED achieves state-of-the-art performance in exemplar-free settings across various scenarios, showing the potential of expert diversification through data in continual learning.

Results

TaskDatasetMetricValueModel
Continual LearningCifar100-B0(10 tasks)-no-exemplarsAverage Incremental Accuracy61.7SEED
Continual LearningCIFAR100-B0(50 tasks)-no-exemplarsAverage Incremental Accuracy42.6SEED
Continual LearningCifar100-B0(20 tasks)-no-exemplarsAverage Incremental Accuracy56.2SEED
Class Incremental LearningCifar100-B0(10 tasks)-no-exemplarsAverage Incremental Accuracy61.7SEED
Class Incremental LearningCIFAR100-B0(50 tasks)-no-exemplarsAverage Incremental Accuracy42.6SEED
Class Incremental LearningCifar100-B0(20 tasks)-no-exemplarsAverage Incremental Accuracy56.2SEED

Related Papers

RegCL: Continual Adaptation of Segment Anything Model via Model Merging2025-07-16Information-Theoretic Generalization Bounds of Replay-based Continual Learning2025-07-16PROL : Rehearsal Free Continual Learning in Streaming Data via Prompt Online Learning2025-07-16Fast Last-Iterate Convergence of SGD in the Smooth Interpolation Regime2025-07-15A Neural Network Model of Complementary Learning Systems: Pattern Separation and Completion for Continual Learning2025-07-15LifelongPR: Lifelong knowledge fusion for point cloud place recognition based on replay and prompt learning2025-07-14Overcoming catastrophic forgetting in neural networks2025-07-14Continual Reinforcement Learning by Planning with Online World Models2025-07-12