Attention-Free Keyword Spotting

14 Oct 2021  ·  Mashrur M. Morshed, Ahmad Omar Ahsan ·

Till now, attention-based models have been used with great success in the keyword spotting problem domain. However, in light of recent advances in deep learning, the question arises whether self-attention is truly irreplaceable for recognizing speech keywords. We thus explore the usage of gated MLPs --previously shown to be alternatives to transformers in vision tasks-- for the keyword spotting task. We provide a family of highly efficient MLP-based models for keyword spotting, with less than 0.5 million parameters. We show that our approach achieves competitive performance on Google Speech Commands V2-12 and V2-35 benchmarks with much fewer parameters than self-attention-based methods.

PDF Abstract

Datasets


Results from the Paper


Ranked #8 on Keyword Spotting on Google Speech Commands (Google Speech Commands V2 35 metric)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Keyword Spotting Google Speech Commands KW-MLP Google Speech Commands V2 35 97.56 # 8

Methods


No methods listed for this paper. Add relevant methods here