Unsupervised Machine Translation

32 papers with code • 9 benchmarks • 4 datasets

Unsupervised machine translation is the task of doing machine translation without any translation resources at training time.

( Image credit: Phrase-Based & Neural Unsupervised Machine Translation )

Libraries

Use these libraries to find Unsupervised Machine Translation models and implementations

Quick Back-Translation for Unsupervised Machine Translation

bbrimacombe/quick-back-translation 1 Dec 2023

The field of unsupervised machine translation has seen significant advancement from the marriage of the Transformer and the back-translation algorithm.

2
01 Dec 2023

Weakly-supervised Deep Cognate Detection Framework for Low-Resourced Languages Using Morphological Knowledge of Closely-Related Languages

koustavagoswami/weakly_supervised-cognate_detection 9 Nov 2023

We train an encoder to gain morphological knowledge of a language and transfer the knowledge to perform unsupervised and weakly-supervised cognate detection tasks with and without the pivot language for the closely-related languages.

1
09 Nov 2023

Bilex Rx: Lexical Data Augmentation for Massively Multilingual Machine Translation

google-research/url-nlp 27 Mar 2023

Neural machine translation (NMT) has progressed rapidly over the past several years, and modern models are able to achieve relatively high quality using only monolingual text data, an approach dubbed Unsupervised Machine Translation (UNMT).

166
27 Mar 2023

Unsupervised Mandarin-Cantonese Machine Translation

meganndare/cantonese-nlp 10 Jan 2023

Advancements in unsupervised machine translation have enabled the development of machine translation systems that can translate between languages for which there is not an abundance of parallel data available.

24
10 Jan 2023

Refining Low-Resource Unsupervised Translation by Language Disentanglement of Multilingual Model

nxphi47/refine_unsup_multilingual_mt 31 May 2022

Numerous recent work on unsupervised machine translation (UMT) implies that competent unsupervised translations of low-resource and unrelated languages, such as Nepali or Sinhala, are only possible if the model is trained in a massive multilingual environment, where these low-resource languages are mixed with high-resource counterparts.

1
31 May 2022

Leveraging Automated Unit Tests for Unsupervised Code Translation

facebookresearch/CodeGen ICLR 2022

With little to no parallel data available for programming languages, unsupervised methods are well-suited to source code translation.

673
13 Oct 2021

Unsupervised Translation of German--Lower Sorbian: Exploring Training and Novel Transfer Methods on a Low-Resource Language

leukas/wmt21 24 Sep 2021

Lastly, we experiment with the order in which offline and online back-translation are used to train an unsupervised system, finding that using online back-translation first works better for DE$\rightarrow$DSB by 2. 76 BLEU.

1
24 Sep 2021

Break-It-Fix-It: Unsupervised Learning for Program Repair

michiyasunaga/bifi 11 Jun 2021

To bridge this gap, we propose a new training approach, Break-It-Fix-It (BIFI), which has two key ideas: (i) we use the critic to check a fixer's output on real bad inputs and add good (fixed) outputs to the training data, and (ii) we train a breaker to generate realistic bad code from good code.

107
11 Jun 2021

Zero-Shot Language Transfer vs Iterative Back Translation for Unsupervised Machine Translation

chanzy3/11747_Final_Project 31 Mar 2021

This work focuses on comparing different solutions for machine translation on low resource language pairs, namely, with zero-shot transfer learning and unsupervised machine translation.

2
31 Mar 2021

The LMU Munich System for the WMT 2020 Unsupervised Machine Translation Shared Task

alexandra-chron/umt-lmu-wmt2020 WMT (EMNLP) 2020

Our core unsupervised neural machine translation (UNMT) system follows the strategy of Chronopoulou et al. (2020), using a monolingual pretrained language generation model (on German) and fine-tuning it on both German and Upper Sorbian, before initializing a UNMT model, which is trained with online backtranslation.

6
25 Oct 2020