TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Unsupervised Statistical Machine Translation

Unsupervised Statistical Machine Translation

Mikel Artetxe, Gorka Labaka, Eneko Agirre

2018-09-04EMNLP 2018 10Machine TranslationNMTUnsupervised Machine TranslationTranslationLanguage Modelling
PaperPDFCodeCode(official)Code(official)

Abstract

While modern machine translation has relied on large parallel corpora, a recent line of work has managed to train Neural Machine Translation (NMT) systems from monolingual corpora only (Artetxe et al., 2018c; Lample et al., 2018). Despite the potential of this approach for low-resource settings, existing systems are far behind their supervised counterparts, limiting their practical interest. In this paper, we propose an alternative approach based on phrase-based Statistical Machine Translation (SMT) that significantly closes the gap with supervised systems. Our method profits from the modular architecture of SMT: we first induce a phrase table from monolingual corpora through cross-lingual embedding mappings, combine it with an n-gram language model, and fine-tune hyperparameters through an unsupervised MERT variant. In addition, iterative backtranslation improves results further, yielding, for instance, 14.08 and 26.22 BLEU points in WMT 2014 English-German and English-French, respectively, an improvement of more than 7-10 BLEU points over previous unsupervised systems, and closing the gap with supervised SMT (Moses trained on Europarl) down to 2-5 BLEU points. Our implementation is available at https://github.com/artetxem/monoses

Results

TaskDatasetMetricValueModel
Machine TranslationWMT2014 French-EnglishBLEU score25.87SMT + iterative backtranslation (unsupervised)
Machine TranslationWMT2016 English-GermanBLEU score18.23SMT + iterative backtranslation (unsupervised)
Machine TranslationWMT2016 German-EnglishBLEU score23.05SMT + iterative backtranslation (unsupervised)
Machine TranslationWMT2014 German-EnglishBLEU score17.43SMT + iterative backtranslation (unsupervised)
Machine TranslationWMT2014 English-GermanBLEU score14.08SMT + iterative backtranslation (unsupervised)
Machine TranslationWMT2014 English-FrenchBLEU score26.22SMT + iterative backtranslation (unsupervised)
Machine TranslationWMT2014 French-EnglishBLEU25.9SMT

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17Making Language Model a Hierarchical Classifier and Generator2025-07-17VisionThink: Smart and Efficient Vision Language Model via Reinforcement Learning2025-07-17The Generative Energy Arena (GEA): Incorporating Energy Awareness in Large Language Model (LLM) Human Evaluations2025-07-17Inverse Reinforcement Learning Meets Large Language Model Post-Training: Basics, Advances, and Opportunities2025-07-17Assay2Mol: large language model-based drug design using BioAssay context2025-07-16Describe Anything Model for Visual Question Answering on Text-rich Images2025-07-16