no code implementations • 29 Feb 2024 • Tsz Kin Lam, Alexandra Birch, Barry Haddow
In this paper, we leverage the SSL models by pretraining smaller models on their Discrete Speech Units (DSU).
no code implementations • 1 Feb 2024 • Giulio Zhou, Tsz Kin Lam, Alexandra Birch, Barry Haddow
While there has been a growing interest in developing direct speech translation systems to avoid propagating errors and losing non-verbal content, prior work in direct S2TT has struggled to conclusively establish the advantages of integrating the acoustic signal directly into the translation process.
no code implementations • 27 Oct 2022 • Tsz Kin Lam, Shigehiko Schamoni, Stefan Riezler
Data augmentation is a technique to generate new training data based on existing data.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 24 Oct 2022 • Tsz Kin Lam, Eva Hasler, Felix Hieber
Customer feedback can be an important signal for improving commercial machine translation systems.
no code implementations • ACL 2022 • Tsz Kin Lam, Shigehiko Schamoni, Stefan Riezler
End-to-end speech translation relies on data that pair source-language speech inputs with corresponding translations into a target language.
1 code implementation • 3 Apr 2021 • Tsz Kin Lam, Mayumi Ohta, Shigehiko Schamoni, Stefan Riezler
Our method, called Aligned Data Augmentation (ADA) for ASR, replaces transcribed tokens and the speech representations in an aligned manner to generate previously unseen training pairs.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • 21 Oct 2020 • Tsz Kin Lam, Shigehiko Schamoni, Stefan Riezler
Direct speech translation describes a scenario where only speech inputs and corresponding translations are available.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • WS 2019 • Tsz Kin Lam, Shigehiko Schamoni, Stefan Riezler
We propose an interactive-predictive neural machine translation framework for easier model personalization using reinforcement and imitation learning.
1 code implementation • 3 May 2018 • Tsz Kin Lam, Julia Kreutzer, Stefan Riezler
We present an approach to interactive-predictive neural machine translation that attempts to reduce human effort from three directions: Firstly, instead of requiring humans to select, correct, or delete segments, we employ the idea of learning from human reinforcements in form of judgments on the quality of partial translations.