TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Towards Lingua Franca Named Entity Recognition with BERT

Towards Lingua Franca Named Entity Recognition with BERT

Taesun Moon, Parul Awasthy, Jian Ni, Radu Florian

2019-11-19named-entity-recognitionCross-Lingual TransferNamed Entity RecognitionCross-Lingual NERNamed Entity Recognition (NER)
PaperPDF

Abstract

Information extraction is an important task in NLP, enabling the automatic extraction of data for relational database filling. Historically, research and data was produced for English text, followed in subsequent years by datasets in Arabic, Chinese (ACE/OntoNotes), Dutch, Spanish, German (CoNLL evaluations), and many others. The natural tendency has been to treat each language as a different dataset and build optimized models for each. In this paper we investigate a single Named Entity Recognition model, based on a multilingual BERT, that is trained jointly on many languages simultaneously, and is able to decode these languages with better accuracy than models trained only on one language. To improve the initial model, we study the use of regularization strategies such as multitask learning and partial gradient updates. In addition to being a single model that can tackle multiple languages (including code switch), the model could be used to make zero-shot predictions on a new language, even ones for which training data is not available, out of the box. The results show that this model not only performs competitively with monolingual models, but it also achieves state-of-the-art results on the CoNLL02 Dutch and Spanish datasets, OntoNotes Arabic and Chinese datasets. Moreover, it performs reasonably well on unseen languages, achieving state-of-the-art for zero-shot on three CoNLL languages.

Results

TaskDatasetMetricValueModel
Cross-LingualCoNLL DutchF183.35Zero shot mBERT 3
Cross-LingualCoNLL GermanF172.44Zero shot mBERT 3
Cross-LingualCoNLL SpanishF176.53Zero shot mBERT 3
Cross-Lingual TransferCoNLL DutchF183.35Zero shot mBERT 3
Cross-Lingual TransferCoNLL GermanF172.44Zero shot mBERT 3
Cross-Lingual TransferCoNLL SpanishF176.53Zero shot mBERT 3

Related Papers

Enhancing Cross-task Transfer of Large Language Models via Activation Steering2025-07-17HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training2025-07-15Flippi: End To End GenAI Assistant for E-Commerce2025-07-08Selecting and Merging: Towards Adaptable and Scalable Named Entity Recognition with Large Language Models2025-06-28Better Semi-supervised Learning for Multi-domain ASR Through Incremental Retraining and Data Filtering2025-06-05Dissecting Bias in LLMs: A Mechanistic Interpretability Perspective2025-06-05Efficient Data Selection for Domain Adaptation of ASR Using Pseudo-Labels and Multi-Stage Filtering2025-06-04Cross-Lingual Transfer of Cultural Knowledge: An Asymmetric Phenomenon2025-06-02