Yuyang Nie, Yuanhe Tian, Yan Song, Xiang Ao, Xiang Wan
Named entity recognition (NER) is highly sensitive to sentential syntactic and semantic properties where entities may be extracted according to how they are used and placed in the running text. To model such properties, one could rely on existing resources to providing helpful knowledge to the NER task; some existing studies proved the effectiveness of doing so, and yet are limited in appropriately leveraging the knowledge such as distinguishing the important ones for particular context. In this paper, we improve NER by leveraging different types of syntactic information through attentive ensemble, which functionalizes by the proposed key-value memory networks, syntax attention, and the gate mechanism for encoding, weighting and aggregating such syntactic information, respectively. Experimental results on six English and Chinese benchmark datasets suggest the effectiveness of the proposed model and show that it outperforms previous studies on all experiment datasets.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Named Entity Recognition (NER) | Ontonotes v5 (English) | F1 | 90.32 | AESINER |
| Named Entity Recognition (NER) | WNUT 2017 | F1 | 50.68 | AESINER |
| Named Entity Recognition (NER) | WNUT 2016 | F1 | 55.14 | AESINER |
| Named Entity Recognition (NER) | Weibo NER | F1 | 69.78 | AESINER |
| Named Entity Recognition (NER) | Resume NER | F1 | 96.62 | AESINER |
| Named Entity Recognition (NER) | OntoNotes 4 | F1 | 81.18 | AESINER |