Search Results for author: Qingyun Dou

Found 4 papers, 1 papers with code

Deliberation Networks and How to Train Them

no code implementations6 Nov 2022 Qingyun Dou, Mark Gales

A deliberation network consists of multiple standard sequence-to-sequence models, each one conditioned on the initial input and the output of the previous model.

Machine Translation Speech Synthesis

Parallel Attention Forcing for Machine Translation

no code implementations6 Nov 2022 Qingyun Dou, Mark Gales

Attention forcing has been introduced to address the mismatch, guiding the model with the generated back-history and reference attention.

Machine Translation NMT +1

Attention Forcing for Machine Translation

1 code implementation2 Apr 2021 Qingyun Dou, Yiting Lu, Potsawee Manakul, Xixin Wu, Mark J. F. Gales

This approach guides the model with the generated output history and reference attention, and can reduce the training-inference mismatch without a schedule or a classifier.

Machine Translation NMT +1

Attention Forcing for Sequence-to-sequence Model Training

no code implementations26 Sep 2019 Qingyun Dou, Yiting Lu, Joshua Efiong, Mark J. F. Gales

This paper introduces attention forcing, which guides the model with generated output history and reference attention.

Machine Translation Speech Synthesis +2

Cannot find the paper you are looking for? You can Submit a new open access paper.