TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Distilling Causal Effect from Miscellaneous Other-Class fo...

Distilling Causal Effect from Miscellaneous Other-Class for Continual Named Entity Recognition

Junhao Zheng, Zhanxian Liang, Haibin Chen, Qianli Ma

2022-10-08Continual Learningnamed-entity-recognitionNamed Entity RecognitionMiscellaneousNERCausal InferenceFG-1-PG-1Named Entity Recognition (NER)
PaperPDFCode(official)

Abstract

Continual Learning for Named Entity Recognition (CL-NER) aims to learn a growing number of entity types over time from a stream of data. However, simply learning Other-Class in the same way as new entity types amplifies the catastrophic forgetting and leads to a substantial performance drop. The main cause behind this is that Other-Class samples usually contain old entity types, and the old knowledge in these Other-Class samples is not preserved properly. Thanks to the causal inference, we identify that the forgetting is caused by the missing causal effect from the old data. To this end, we propose a unified causal framework to retrieve the causality from both new entity types and Other-Class. Furthermore, we apply curriculum learning to mitigate the impact of label noise and introduce a self-adaptive weight for balancing the causal effects between new entity types and Other-Class. Experimental results on three benchmark datasets show that our method outperforms the state-of-the-art method by a large margin. Moreover, our method can be combined with the existing state-of-the-art methods to improve the performance in CL-NER

Results

TaskDatasetMetricValueModel
Continual LearningOntoNotes 5.0F1 (macro)0.4222CFNER
Continual LearningOntoNotes 5.0F1 (micro)0.5894CFNER
Continual Learning2010 i2b2/VAF1 (macro)0.3626CFNER
Continual Learning2010 i2b2/VAF1 (micro)0.6273CFNER
Continual Learningconll2003F1 (macro)0.7911CFNER
Continual Learningconll2003F1 (micro)0.8091CFNER
Continual Named Entity RecognitionOntoNotes 5.0F1 (macro)0.4222CFNER
Continual Named Entity RecognitionOntoNotes 5.0F1 (micro)0.5894CFNER
Continual Named Entity Recognition2010 i2b2/VAF1 (macro)0.3626CFNER
Continual Named Entity Recognition2010 i2b2/VAF1 (micro)0.6273CFNER
Continual Named Entity Recognitionconll2003F1 (macro)0.7911CFNER
Continual Named Entity Recognitionconll2003F1 (micro)0.8091CFNER

Related Papers

RegCL: Continual Adaptation of Segment Anything Model via Model Merging2025-07-16Information-Theoretic Generalization Bounds of Replay-based Continual Learning2025-07-16PROL : Rehearsal Free Continual Learning in Streaming Data via Prompt Online Learning2025-07-16Fast Last-Iterate Convergence of SGD in the Smooth Interpolation Regime2025-07-15A Neural Network Model of Complementary Learning Systems: Pattern Separation and Completion for Continual Learning2025-07-15LifelongPR: Lifelong knowledge fusion for point cloud place recognition based on replay and prompt learning2025-07-14Overcoming catastrophic forgetting in neural networks2025-07-14Continual Reinforcement Learning by Planning with Online World Models2025-07-12