Understanding Back-Translation at Scale

EMNLP 2018 Sergey EdunovMyle OttMichael AuliDavid Grangier

An effective method to improve neural machine translation with monolingual data is to augment the parallel training corpus with back-translations of target language sentences. This work broadens the understanding of back-translation and investigates a number of methods to generate synthetic source sentences... (read more)

PDF Abstract

Results from the Paper


 Ranked #1 on Machine Translation on WMT2014 English-German (using extra training data)

     Get a GitHub badge
TASK DATASET MODEL METRIC NAME METRIC VALUE GLOBAL RANK USES EXTRA
TRAINING DATA
RESULT BENCHMARK
Machine Translation WMT2014 English-French Transformer Big + BT BLEU score 45.6 # 1
SacreBLEU 43.8 # 1
Machine Translation WMT2014 English-German Transformer Big + BT BLEU score 35.0 # 1
SacreBLEU 33.8 # 1

Methods used in the Paper


METHOD TYPE
🤖 No Methods Found Help the community by adding them if they're not listed; e.g. Deep Residual Learning for Image Recognition uses ResNet