Weak-Attention Suppression For Transformer Based Speech Recognition

18 May 2020Yangyang ShiYongqiang WangChunyang WuChristian FuegenFrank ZhangDuc LeChing-Feng YehMichael L. Seltzer

Transformers, originally proposed for natural language processing (NLP) tasks, have recently achieved great success in automatic speech recognition (ASR). However, adjacent acoustic units (i.e., frames) are highly correlated, and long-distance dependencies between them are weak, unlike text units... (read more)

PDF Abstract


No code implementations yet. Submit your code now

Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper