TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/PromptNER: Prompt Locating and Typing for Named Entity Rec...

PromptNER: Prompt Locating and Typing for Named Entity Recognition

Yongliang Shen, Zeqi Tan, Shuhui Wu, Wenqi Zhang, Rongsheng Zhang, Yadong Xi, Weiming Lu, Yueting Zhuang

2023-05-26Nested Named Entity Recognitionnamed-entity-recognitionNamed Entity RecognitionCross-Domain Few-ShotNERGraph MatchingNamed Entity Recognition (NER)Entity Typing
PaperPDFCode(official)

Abstract

Prompt learning is a new paradigm for utilizing pre-trained language models and has achieved great success in many tasks. To adopt prompt learning in the NER task, two kinds of methods have been explored from a pair of symmetric perspectives, populating the template by enumerating spans to predict their entity types or constructing type-specific prompts to locate entities. However, these methods not only require a multi-round prompting manner with a high time overhead and computational cost, but also require elaborate prompt templates, that are difficult to apply in practical scenarios. In this paper, we unify entity locating and entity typing into prompt learning, and design a dual-slot multi-prompt template with the position slot and type slot to prompt locating and typing respectively. Multiple prompts can be input to the model simultaneously, and then the model extracts all entities by parallel predictions on the slots. To assign labels for the slots during training, we design a dynamic template filling mechanism that uses the extended bipartite graph matching between prompts and the ground-truth entities. We conduct experiments in various settings, including resource-rich flat and nested NER datasets and low-resource in-domain and cross-domain datasets. Experimental results show that the proposed model achieves a significant performance improvement, especially in the cross-domain few-shot setting, which outperforms the state-of-the-art model by +7.7% on average.

Results

TaskDatasetMetricValueModel
Named Entity Recognition (NER)ACE 2005F188.26PromptNER [RoBERTa-large]
Named Entity Recognition (NER)ACE 2005F187.21PromptNER [BERT-large]
Named Entity Recognition (NER)CoNLL 2003 (English)F193.08PromptNER [RoBERTa-large]
Named Entity Recognition (NER)CoNLL 2003 (English)F192.41PromptNER [BERT-large]
Named Entity Recognition (NER)ACE 2004F188.72PromptNER [RoBERTa-large]
Named Entity Recognition (NER)ACE 2004F188.16PromptNER [BERT-large]

Related Papers

Flippi: End To End GenAI Assistant for E-Commerce2025-07-08Selecting and Merging: Towards Adaptable and Scalable Named Entity Recognition with Large Language Models2025-06-28Improving Named Entity Transcription with Contextual LLM-based Revision2025-06-12Adapter Naturally Serves as Decoupler for Cross-Domain Few-Shot Semantic Segmentation2025-06-09Better Semi-supervised Learning for Multi-domain ASR Through Incremental Retraining and Data Filtering2025-06-05Dissecting Bias in LLMs: A Mechanistic Interpretability Perspective2025-06-05Efficient Data Selection for Domain Adaptation of ASR Using Pseudo-Labels and Multi-Stage Filtering2025-06-04Probing Neural Topology of Large Language Models2025-06-01