Search Results for author: Afra Alishahi

Found 21 papers, 12 papers with code

Encoding of lexical tone in self-supervised models of spoken language

no code implementations25 Mar 2024 Gaofei Shen, Michaela Watkins, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała

Interpretability research has shown that self-supervised Spoken Language Models (SLMs) encode a wide variety of features in human speech from the acoustic, phonetic, phonological, syntactic and semantic levels, to speaker characteristics.

Homophone Disambiguation Reveals Patterns of Context Mixing in Speech Transformers

1 code implementation15 Oct 2023 Hosein Mohebbi, Grzegorz Chrupała, Willem Zuidema, Afra Alishahi

Transformers have become a key architecture in speech processing, but our understanding of how they build up representations of acoustic and linguistic structure is limited.

Decoder speech-recognition +1

Wave to Syntax: Probing spoken language models for syntax

1 code implementation30 May 2023 Gaofei Shen, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała

Understanding which information is encoded in deep models of spoken and written language has been the focus of much research in recent years, as it is crucial for debugging and improving these architectures.

Quantifying Context Mixing in Transformers

1 code implementation30 Jan 2023 Hosein Mohebbi, Willem Zuidema, Grzegorz Chrupała, Afra Alishahi

Self-attention weights and their transformed variants have been the main source of information for analyzing token-to-token interactions in Transformer-based models.

Learning English with Peppa Pig

1 code implementation25 Feb 2022 Mitja Nikolaus, Afra Alishahi, Grzegorz Chrupała

In the real world the coupling between the linguistic and the visual modality is loose, and often confounded by correlations with non-semantic aspects of the speech signal.

Descriptive

Discrete representations in neural models of spoken language

1 code implementation EMNLP (BlackboxNLP) 2021 Bertrand Higy, Lieke Gelderloos, Afra Alishahi, Grzegorz Chrupała

The distributed and continuous representations used by neural networks are at odds with representations employed in linguistics, which are typically symbolic.

Attribute Quantization

Learning to Understand Child-directed and Adult-directed Speech

no code implementations ACL 2020 Lieke Gelderloos, Grzegorz Chrupała, Afra Alishahi

Speech directed to children differs from adult-directed speech in linguistic aspects such as repetition, word choice, and sentence length, as well as in aspects of the speech signal itself, such as prosodic and phonemic variation.

Language Acquisition Sentence

Analyzing analytical methods: The case of phonology in neural models of spoken language

1 code implementation ACL 2020 Grzegorz Chrupała, Bertrand Higy, Afra Alishahi

Given the fast development of analysis techniques for NLP and speech processing systems, few systematic studies have been conducted to compare the strengths and weaknesses of each method.

Bootstrapping Disjoint Datasets for Multilingual Multimodal Representation Learning

no code implementations9 Nov 2019 Ákos Kádár, Grzegorz Chrupała, Afra Alishahi, Desmond Elliott

However, we do find that using an external machine translation model to generate the synthetic data sets results in better performance.

Machine Translation Representation Learning +4

Correlating neural and symbolic representations of language

1 code implementation ACL 2019 Grzegorz Chrupała, Afra Alishahi

Analysis methods which enable us to better understand the representations and functioning of neural models of language are increasingly needed as deep learning becomes the dominant approach in NLP.

Analyzing and Interpreting Neural Networks for NLP: A Report on the First BlackboxNLP Workshop

no code implementations5 Apr 2019 Afra Alishahi, Grzegorz Chrupała, Tal Linzen

The EMNLP 2018 workshop BlackboxNLP was dedicated to resources and techniques specifically developed for analyzing and understanding the inner-workings and representations acquired by neural models of language.

Revisiting the Hierarchical Multiscale LSTM

no code implementations COLING 2018 Ákos Kádár, Marc-Alexandre Côté, Grzegorz Chrupała, Afra Alishahi

Hierarchical Multiscale LSTM (Chung et al., 2016a) is a state-of-the-art language model that learns interpretable structure from character-level input.

Language Modelling

On the difficulty of a distributional semantics of spoken language

no code implementations WS 2019 Grzegorz Chrupała, Lieke Gelderloos, Ákos Kádár, Afra Alishahi

In the domain of unsupervised learning most work on speech has focused on discovering low-level constructs such as phoneme inventories or word-like units.

Encoding of phonology in a recurrent neural model of grounded speech

1 code implementation CONLL 2017 Afra Alishahi, Marie Barking, Grzegorz Chrupała

We study the representation and encoding of phonemes in a recurrent neural network model of grounded speech.

Clustering

Representations of language in a model of visually grounded speech signal

2 code implementations ACL 2017 Grzegorz Chrupała, Lieke Gelderloos, Afra Alishahi

We present a visually grounded model of speech perception which projects spoken utterances and images to a joint semantic space.

Representation of linguistic form and function in recurrent neural networks

1 code implementation CL 2017 Ákos Kádár, Grzegorz Chrupała, Afra Alishahi

We present novel methods for analyzing the activation patterns of RNNs from a linguistic point of view and explore the types of linguistic structure they learn.

Language Modelling Sentence +1

Learning language through pictures

1 code implementation IJCNLP 2015 Grzegorz Chrupała, Ákos Kádár, Afra Alishahi

We propose Imaginet, a model of learning visually grounded representations of language from coupled textual and visual input.

Sentence Word Embeddings

Cannot find the paper you are looking for? You can Submit a new open access paper.