no code implementations • 12 Dec 2023 • Mohammed Maqsood Shaik, Dietrich Klakow, Badr M. Abdullah
To address this challenge, we propose self-supervised adaptive pre-training (SAPT) to adapt the pre-trained model to the target domain and languages of the downstream task.
1 code implementation • 4 Jun 2023 • Badr M. Abdullah, Mohammed Maqsood Shaik, Bernd Möbius, Dietrich Klakow
Self-supervised representation learning for speech often involves a quantization step that transforms the acoustic input into discrete units.
no code implementations • 3 Mar 2023 • Kanisius Karyono, Badr M. Abdullah, Alison J. Cotgrave, Ana Bras, Jeff Cullen
This work introduces the reliable data set for training the AI subsystem for thermal comfort.
no code implementations • 8 Jan 2023 • Badr M. Abdullah, Dietrich Klakow
In this paper, we take a closer analytical look at AWEs learned from English speech and study how the choice of the learning objective and the architecture shapes their representational profile.
1 code implementation • 14 Sep 2022 • Badr M. Abdullah, Bernd Möbius, Dietrich Klakow
Models of acoustic word embeddings (AWEs) learn to map variable-length spoken word segments onto fixed-dimensionality vector representations such that different acoustic exemplars of the same word are projected nearby in the embedding space.
1 code implementation • EMNLP (BlackboxNLP) 2021 • Badr M. Abdullah, Iuliia Zaitova, Tania Avgustinova, Bernd Möbius, Dietrich Klakow
We further discuss the implications of our work on modeling speech processing and language similarity with neural networks.
1 code implementation • 16 Jun 2021 • Badr M. Abdullah, Marius Mosbach, Iuliia Zaitova, Bernd Möbius, Dietrich Klakow
Our experiments show that (1) the distance in the embedding space in the best cases only moderately correlates with phonological distance, and (2) improving the performance on the word discrimination task does not necessarily yield models that better reflect word phonological similarity.
no code implementations • NAACL (SIGTYP) 2021 • Elizabeth Salesky, Badr M. Abdullah, Sabrina J. Mielke, Elena Klyachko, Oleg Serikov, Edoardo Ponti, Ritesh Kumar, Ryan Cotterell, Ekaterina Vylomova
While language identification is a fundamental speech and language processing task, for many languages and language families it remains a challenging task.
no code implementations • EACL 2021 • Nicole Macher, Badr M. Abdullah, Harm Brouwer, Dietrich Klakow
Theories and models of spoken word recognition aim to explain the process of accessing lexical knowledge given an acoustic realization of a word form.
no code implementations • EACL 2021 • Alexandra Mayn, Badr M. Abdullah, Dietrich Klakow
We present a deep neural model of spoken word recognition which is trained to retrieve the meaning of a word (in the form of a word embedding) given its spoken form, a task which resembles that faced by a human listener.
1 code implementation • COLING 2020 • Marius Mosbach, Stefania Degaetano-Ortlieb, Marie-Pauline Krielke, Badr M. Abdullah, Dietrich Klakow
Transformer-based language models achieve high performance on various tasks, but we still lack understanding of the kind of linguistic knowledge they learn and rely on.
no code implementations • VarDial (COLING) 2020 • Badr M. Abdullah, Jacek Kudera, Tania Avgustinova, Bernd Möbius, Dietrich Klakow
In this paper, we present a neural model for Slavic language identification in speech signals and analyze its emergent representations to investigate whether they reflect objective measures of language relatedness and/or non-linguists' perception of language similarity.
1 code implementation • 2 Aug 2020 • Badr M. Abdullah, Tania Avgustinova, Bernd Möbius, Dietrich Klakow
State-of-the-art spoken language identification (LID) systems, which are based on end-to-end deep neural networks, have shown remarkable success not only in discriminating between distant languages but also between closely-related languages or even different spoken varieties of the same language.