Unsupervised Machine Translation

32 papers with code • 9 benchmarks • 4 datasets

Unsupervised machine translation is the task of doing machine translation without any translation resources at training time.

( Image credit: Phrase-Based & Neural Unsupervised Machine Translation )

Libraries

Use these libraries to find Unsupervised Machine Translation models and implementations

Unsupervised Vision-and-Language Pre-training Without Parallel Images and Captions

uclanlp/visualbert NAACL 2021

Pre-trained contextual vision-and-language (V&L) models have achieved impressive performance on various benchmarks.

516
24 Oct 2020

A Retrieve-and-Rewrite Initialization Method for Unsupervised Machine Translation

Imagist-Shuo/RRforUNMT ACL 2020

The commonly used framework for unsupervised machine translation builds initial translation models of both translation directions, and then performs iterative back-translation to jointly boost their translation performance.

4
01 Jul 2020

Cross-lingual Retrieval for Iterative Self-Supervised Training

pytorch/fairseq NeurIPS 2020

Recent studies have demonstrated the cross-lingual alignment ability of multilingual pretrained language models.

29,301
16 Jun 2020

Unsupervised Translation of Programming Languages

facebookresearch/TransCoder NeurIPS 2020

We train our model on source code from open source GitHub projects, and show that it can translate functions between C++, Java, and Python with high accuracy.

1,666
05 Jun 2020

Cross-model Back-translated Distillation for Unsupervised Machine Translation

nxphi47/multiagent_crosstranslate 3 Jun 2020

Recent unsupervised machine translation (UMT) systems usually employ three main principles: initialization, language modeling and iterative back-translation, though they may apply them differently.

5
03 Jun 2020

Language Models are Few-Shot Learners

ggerganov/llama.cpp NeurIPS 2020

By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do.

57,427
28 May 2020

Incorporating BERT into Neural Machine Translation

bert-nmt/bert-nmt ICLR 2020

While BERT is more commonly used as fine-tuning instead of contextual embedding for downstream language understanding tasks, in NMT, our preliminary exploration of using BERT as contextual embedding is better than using for fine-tuning.

351
17 Feb 2020

A Probabilistic Formulation of Unsupervised Text Style Transfer

martiansideofthemoon/style-transfer-paraphrase ICLR 2020

Across all style transfer tasks, our approach yields substantial gains over state-of-the-art non-generative baselines, including the state-of-the-art unsupervised machine translation techniques that our approach generalizes.

222
10 Feb 2020

Unsupervised Multilingual Alignment using Wasserstein Barycenter

alixxxin/multi-lang 28 Jan 2020

We study unsupervised multilingual alignment, the problem of finding word-to-word translations between multiple languages without using any parallel data.

8
28 Jan 2020

Multilingual Denoising Pre-training for Neural Machine Translation

huggingface/transformers 22 Jan 2020

This paper demonstrates that multilingual denoising pre-training produces significant performance gains across a wide variety of machine translation (MT) tasks.

125,425
22 Jan 2020