TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Does BERT Make Any Sense? Interpretable Word Sense Disambi...

Does BERT Make Any Sense? Interpretable Word Sense Disambiguation with Contextualized Embeddings

Gregor Wiedemann, Steffen Remus, Avi Chawla, Chris Biemann

2019-09-23Translationtext-classificationGeneral ClassificationWord Sense Disambiguation
PaperPDFCode(official)

Abstract

Contextualized word embeddings (CWE) such as provided by ELMo (Peters et al., 2018), Flair NLP (Akbik et al., 2018), or BERT (Devlin et al., 2019) are a major recent innovation in NLP. CWEs provide semantic vector representations of words depending on their respective context. Their advantage over static word embeddings has been shown for a number of tasks, such as text classification, sequence tagging, or machine translation. Since vectors of the same word type can vary depending on the respective context, they implicitly provide a model for word sense disambiguation (WSD). We introduce a simple but effective approach to WSD using a nearest neighbor classification on CWEs. We compare the performance of different CWE models for the task and can report improvements above the current state of the art for two standard WSD benchmark datasets. We further show that the pre-trained BERT model is able to place polysemic words into distinct 'sense' regions of the embedding space, while ELMo and Flair NLP do not seem to possess this ability.

Results

TaskDatasetMetricValueModel
Word Sense DisambiguationSensEval 2 Lexical SampleF176.52kNN-BERT
Word Sense DisambiguationSensEval 3 Lexical SampleF180.12kNN-BERT
Word Sense DisambiguationSemEval 2007 Task 7F185.32kNN-BERT + POS (training corpus: WNGT)
Word Sense DisambiguationSemEval 2007 Task 7F181.2kNN-BERT
Word Sense DisambiguationSemEval 2007 Task 17F163.17kNN-BERT + POS (training corpus: SemCor)
Word Sense DisambiguationSemEval 2007 Task 17F160.94kNN-BERT

Related Papers

A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17Function-to-Style Guidance of LLMs for Code Translation2025-07-15GNN-CNN: An Efficient Hybrid Model of Convolutional and Graph Neural Networks for Text Representation2025-07-10Speak2Sign3D: A Multi-modal Pipeline for English Speech to American Sign Language Animation2025-07-09Pun Intended: Multi-Agent Translation of Wordplay with Contrastive Learning and Phonetic-Semantic Embeddings2025-07-09Unconditional Diffusion for Generative Sequential Recommendation2025-07-08GRAFT: A Graph-based Flow-aware Agentic Framework for Document-level Machine Translation2025-07-04