TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/FRAGE: Frequency-Agnostic Word Representation

FRAGE: Frequency-Agnostic Word Representation

Chengyue Gong, Di He, Xu Tan, Tao Qin, Li-Wei Wang, Tie-Yan Liu

2018-09-18NeurIPS 2018 12Text ClassificationMachine TranslationWord SimilarityTranslationWord Embeddingstext-classificationLanguage Modelling
PaperPDFCode(official)Code

Abstract

Continuous word representation (aka word embedding) is a basic building block in many neural network-based models used in natural language processing tasks. Although it is widely accepted that words with similar semantics should be close to each other in the embedding space, we find that word embeddings learned in several tasks are biased towards word frequency: the embeddings of high-frequency and low-frequency words lie in different subregions of the embedding space, and the embedding of a rare word and a popular word can be far from each other even if they are semantically similar. This makes learned word embeddings ineffective, especially for rare words, and consequently limits the performance of these neural network models. In this paper, we develop a neat, simple yet effective way to learn \emph{FRequency-AGnostic word Embedding} (FRAGE) using adversarial training. We conducted comprehensive studies on ten datasets across four natural language processing tasks, including word similarity, language modeling, machine translation and text classification. Results show that with FRAGE, we achieve higher performance than the baselines in all tasks.

Results

TaskDatasetMetricValueModel
Machine TranslationIWSLT2015 German-EnglishBLEU score33.97Transformer with FRAGE
Machine TranslationWMT2014 English-GermanBLEU score29.11Transformer Big with FRAGE
Language ModellingPenn Treebank (Word Level)Test perplexity46.54FRAGE + AWD-LSTM-MoS + dynamic eval
Language ModellingPenn Treebank (Word Level)Validation perplexity47.38FRAGE + AWD-LSTM-MoS + dynamic eval
Language ModellingWikiText-2Test perplexity39.14FRAGE + AWD-LSTM-MoS + dynamic eval
Language ModellingWikiText-2Validation perplexity40.85FRAGE + AWD-LSTM-MoS + dynamic eval

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21Making Language Model a Hierarchical Classifier and Generator2025-07-17A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Assay2Mol: large language model-based drug design using BioAssay context2025-07-16Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16