TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/ELMo

ELMo

Natural Language ProcessingIntroduced 2000234 papers
Source Paper

Description

Embeddings from Language Models, or ELMo, is a type of deep contextualized word representation that models both (1) complex characteristics of word use (e.g., syntax and semantics), and (2) how these uses vary across linguistic contexts (i.e., to model polysemy). Word vectors are learned functions of the internal states of a deep bidirectional language model (biLM), which is pre-trained on a large text corpus.

A biLM combines both a forward and backward LM. ELMo jointly maximizes the log likelihood of the forward and backward directions. To add ELMo to a supervised model, we freeze the weights of the biLM and then concatenate the ELMo vector ELMOktask\textbf{ELMO}^{task}_kELMOktask​ with xk\textbf{x}_kxk​ and pass the ELMO enhanced representation [xk;ELMOktask][\textbf{x}_k; \textbf{ELMO}^{task}_k][xk​;ELMOktask​] into the task RNN. Here xk\textbf{x}_kxk​ is a context-independent token representation for each token position.

Image Source: here

Papers Using This Method

A Comparative Analysis of Static Word Embeddings for Hungarian2025-05-12Embedding-based Approaches to Hyperpartisan News Detection2025-01-02Generative Pretrained Embedding and Hierarchical Irregular Time Series Representation for Daily Living Activity Recognition2024-12-27LLMs are Also Effective Embedding Models: An In-depth Overview2024-12-17From Word Vectors to Multimodal Embeddings: Techniques, Applications, and Future Directions For Large Language Models2024-11-06ELMO: Enhanced Real-time LiDAR Motion Capture through Upsampling2024-10-09Evaluating the Efficacy of AI Techniques in Textual Anonymization: A Comparative Study2024-05-09Where exactly does contextualization in a PLM happen?2023-12-11Semantic Change Detection for the Romanian Language2023-08-23PEvoLM: Protein Sequence Evolutionary Information Language Model2023-08-16On "Scientific Debt" in NLP: A Case for More Rigour in Language Model Pre-Training Research2023-06-05Analyzing the Generalizability of Deep Contextualized Language Representations For Text Classification2023-03-22Classifying Text-Based Conspiracy Tweets related to COVID-19 using Contextualized Word Embeddings2023-03-07CKG: Dynamic Representation Based on Context and Knowledge Graph2022-12-09A Context-Sensitive Word Embedding Approach for The Detection of Troll Tweets2022-07-17Always Keep your Target in Mind: Studying Semantics and Improving Performance of Neural Lexical Substitution2022-06-07Parameter-Efficient Tuning by Manipulating Hidden States of Pretrained Language Models For Classification Tasks2022-04-10Bridging Pre-trained Language Models and Hand-crafted Features for Unsupervised POS Tagging2022-03-19Using Word Embeddings to Analyze Protests News2022-03-11Assessment of contextualised representations in detecting outcome phrases in clinical trials2022-02-13