TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Why Attention? Analyze BiLSTM Deficiency and Its Remedies ...

Why Attention? Analyze BiLSTM Deficiency and Its Remedies in the Case of NER

Peng-Hsuan Li, Tsu-Jui Fu, Wei-Yun Ma

2019-08-29NERNamed Entity Recognition (NER)
PaperPDFCode(official)Code(official)Code

Abstract

BiLSTM has been prevalently used as a core module for NER in a sequence-labeling setup. State-of-the-art approaches use BiLSTM with additional resources such as gazetteers, language-modeling, or multi-task supervision to further improve NER. This paper instead takes a step back and focuses on analyzing problems of BiLSTM itself and how exactly self-attention can bring improvements. We formally show the limitation of (CRF-)BiLSTM in modeling cross-context patterns for each word -- the XOR limitation. Then, we show that two types of simple cross-structures -- self-attention and Cross-BiLSTM -- can effectively remedy the problem. We test the practical impacts of the deficiency on real-world NER datasets, OntoNotes 5.0 and WNUT 2017, with clear and consistent improvements over the baseline, up to 8.7% on some of the multi-token entity mentions. We give in-depth analyses of the improvements across several aspects of NER, especially the identification of multi-token mentions. This study should lay a sound foundation for future improvements on sequence-labeling NER. (Source codes: https://github.com/jacobvsdanniel/cross-ner)

Results

TaskDatasetMetricValueModel
Named Entity Recognition (NER)Ontonotes v5 (English)F188.4Att-BiLSTM-CNN
Named Entity Recognition (NER)Ontonotes v5 (English)Precision88.71Att-BiLSTM-CNN
Named Entity Recognition (NER)Ontonotes v5 (English)Recall88.11Att-BiLSTM-CNN
Named Entity Recognition (NER)WNUT 2017F142.85Cross-BiLSTM-CNN
Named Entity Recognition (NER)WNUT 2017Precision58.28Cross-BiLSTM-CNN
Named Entity Recognition (NER)WNUT 2017Recall33.92Cross-BiLSTM-CNN

Related Papers

Flippi: End To End GenAI Assistant for E-Commerce2025-07-08Selecting and Merging: Towards Adaptable and Scalable Named Entity Recognition with Large Language Models2025-06-28Improving Named Entity Transcription with Contextual LLM-based Revision2025-06-12Better Semi-supervised Learning for Multi-domain ASR Through Incremental Retraining and Data Filtering2025-06-05Efficient Data Selection for Domain Adaptation of ASR Using Pseudo-Labels and Multi-Stage Filtering2025-06-04EL4NER: Ensemble Learning for Named Entity Recognition via Multiple Small-Parameter Large Language Models2025-05-29Label-Guided In-Context Learning for Named Entity Recognition2025-05-29Named Entity Recognition in Historical Italian: The Case of Giacomo Leopardi's Zibaldone2025-05-26