no code implementations • 27 Oct 2022 • Chak-Fai Li, Francis Keith, William Hartmann, Matthew Snover
Advances in self-supervised learning have significantly reduced the amount of transcribed audio required for training.
no code implementations • 29 Oct 2021 • Chak-Fai Li, Francis Keith, William Hartmann, Matthew Snover
Final performance is an additional 2% better absolute when using CTC-based decoding for semi-supervised training compared to shallow fusion.
no code implementations • 14 Jun 2021 • Chak-Fai Li, Francis Keith, William Hartmann, Matthew Snover, Owen Kimball
We show that there is a sizable initial gap in such a data condition between hybrid and seq2seq models, and the hybrid model is able to further improve through the use of additional language model (LM) data.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3