TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/UniTrans: Unifying Model Transfer and Data Transfer for Cr...

UniTrans: Unifying Model Transfer and Data Transfer for Cross-Lingual Named Entity Recognition with Unlabeled Data

Qianhui Wu, Zijia Lin, Börje F. Karlsson, Biqing Huang, Jian-Guang Lou

2020-07-15named-entity-recognitionCross-Lingual TransferNamed Entity RecognitionTranslationNERCross-Lingual NERKnowledge DistillationNamed Entity Recognition (NER)
PaperPDFCode(official)

Abstract

Prior works in cross-lingual named entity recognition (NER) with no/little labeled data fall into two primary categories: model transfer based and data transfer based methods. In this paper we find that both method types can complement each other, in the sense that, the former can exploit context information via language-independent features but sees no task-specific information in the target language; while the latter generally generates pseudo target-language training data via translation but its exploitation of context information is weakened by inaccurate translations. Moreover, prior works rarely leverage unlabeled data in the target language, which can be effortlessly collected and potentially contains valuable information for improved results. To handle both problems, we propose a novel approach termed UniTrans to Unify both model and data Transfer for cross-lingual NER, and furthermore, to leverage the available information from unlabeled target-language data via enhanced knowledge distillation. We evaluate our proposed UniTrans over 4 target languages on benchmark datasets. Our experimental results show that it substantially outperforms the existing state-of-the-art methods.

Results

TaskDatasetMetricValueModel
Cross-LingualCoNLL DutchF182.9UniTrans
Cross-LingualCoNLL GermanF174.82UniTrans
Cross-LingualCoNLL SpanishF179.31UniTrans
Cross-LingualNoDaLiDa Norwegian BokmålF181.17UniTrans
Cross-Lingual TransferCoNLL DutchF182.9UniTrans
Cross-Lingual TransferCoNLL GermanF174.82UniTrans
Cross-Lingual TransferCoNLL SpanishF179.31UniTrans
Cross-Lingual TransferNoDaLiDa Norwegian BokmålF181.17UniTrans

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Enhancing Cross-task Transfer of Large Language Models via Activation Steering2025-07-17A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training2025-07-15Function-to-Style Guidance of LLMs for Code Translation2025-07-15Feature Distillation is the Better Choice for Model-Heterogeneous Federated Learning2025-07-14