Language Models

mBART is a sequence-to-sequence denoising auto-encoder pre-trained on large-scale monolingual corpora in many languages using the BART objective. The input texts are noised by masking phrases and permuting sentences, and a single Transformer model is learned to recover the texts. Different from other pre-training approaches for machine translation, mBART pre-trains a complete autoregressive Seq2Seq model. mBART is trained once for all languages, providing a set of parameters that can be fine-tuned for any of the language pairs in both supervised and unsupervised settings, without any task-specific or language-specific modifications or initialization schemes.

Source: Multilingual Denoising Pre-training for Neural Machine Translation

Papers


Paper Code Results Date Stars

Tasks


Task Papers Share
Translation 36 20.93%
Machine Translation 27 15.70%
Sentence 11 6.40%
Text Generation 8 4.65%
Denoising 8 4.65%
NMT 7 4.07%
Abstractive Text Summarization 6 3.49%
Text Summarization 5 2.91%
Cross-Lingual Transfer 5 2.91%

Components


Component Type
🤖 No Components Found You can add them if they exist; e.g. Mask R-CNN uses RoIAlign

Categories