Search Results for author: Grzegorz Chrupała

Found 29 papers, 19 papers with code

Encoding of lexical tone in self-supervised models of spoken language

no code implementations25 Mar 2024 Gaofei Shen, Michaela Watkins, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała

Interpretability research has shown that self-supervised Spoken Language Models (SLMs) encode a wide variety of features in human speech from the acoustic, phonetic, phonological, syntactic and semantic levels, to speaker characteristics.

Homophone Disambiguation Reveals Patterns of Context Mixing in Speech Transformers

1 code implementation15 Oct 2023 Hosein Mohebbi, Grzegorz Chrupała, Willem Zuidema, Afra Alishahi

Transformers have become a key architecture in speech processing, but our understanding of how they build up representations of acoustic and linguistic structure is limited.

Decoder speech-recognition +1

Quantifying the Plausibility of Context Reliance in Neural Machine Translation

2 code implementations2 Oct 2023 Gabriele Sarti, Grzegorz Chrupała, Malvina Nissim, Arianna Bisazza

Establishing whether language models can use contextual information in a human-plausible way is important to ensure their trustworthiness in real-world settings.

Machine Translation Translation

Wave to Syntax: Probing spoken language models for syntax

1 code implementation30 May 2023 Gaofei Shen, Afra Alishahi, Arianna Bisazza, Grzegorz Chrupała

Understanding which information is encoded in deep models of spoken and written language has been the focus of much research in recent years, as it is crucial for debugging and improving these architectures.

Putting Natural in Natural Language Processing

no code implementations8 May 2023 Grzegorz Chrupała

Human language is firstly spoken and only secondarily written.

Quantifying Context Mixing in Transformers

1 code implementation30 Jan 2023 Hosein Mohebbi, Willem Zuidema, Grzegorz Chrupała, Afra Alishahi

Self-attention weights and their transformed variants have been the main source of information for analyzing token-to-token interactions in Transformer-based models.

Learning English with Peppa Pig

1 code implementation25 Feb 2022 Mitja Nikolaus, Afra Alishahi, Grzegorz Chrupała

In the real world the coupling between the linguistic and the visual modality is loose, and often confounded by correlations with non-semantic aspects of the speech signal.

Descriptive

Cyberbullying Classifiers are Sensitive to Model-Agnostic Perturbations

1 code implementation LREC 2022 Chris Emmery, Ákos Kádár, Grzegorz Chrupała, Walter Daelemans

The perturbed data, models, and code are available for reproduction at https://github. com/cmry/augtox

Discrete representations in neural models of spoken language

1 code implementation EMNLP (BlackboxNLP) 2021 Bertrand Higy, Lieke Gelderloos, Afra Alishahi, Grzegorz Chrupała

The distributed and continuous representations used by neural networks are at odds with representations employed in linguistics, which are typically symbolic.

Attribute Quantization

Visually grounded models of spoken language: A survey of datasets, architectures and evaluation techniques

no code implementations27 Apr 2021 Grzegorz Chrupała

This survey provides an overview of the evolution of visually grounded models of spoken language over the last 20 years.

Adversarial Stylometry in the Wild: Transferable Lexical Substitution Attacks on Author Profiling

1 code implementation EACL 2021 Chris Emmery, Ákos Kádár, Grzegorz Chrupała

Written language contains stylistic cues that can be exploited to automatically infer a variety of potentially sensitive author information.

Privacy Preserving

Learning to Understand Child-directed and Adult-directed Speech

no code implementations ACL 2020 Lieke Gelderloos, Grzegorz Chrupała, Afra Alishahi

Speech directed to children differs from adult-directed speech in linguistic aspects such as repetition, word choice, and sentence length, as well as in aspects of the speech signal itself, such as prosodic and phonemic variation.

Language Acquisition Sentence

Analyzing analytical methods: The case of phonology in neural models of spoken language

1 code implementation ACL 2020 Grzegorz Chrupała, Bertrand Higy, Afra Alishahi

Given the fast development of analysis techniques for NLP and speech processing systems, few systematic studies have been conducted to compare the strengths and weaknesses of each method.

Bootstrapping Disjoint Datasets for Multilingual Multimodal Representation Learning

no code implementations9 Nov 2019 Ákos Kádár, Grzegorz Chrupała, Afra Alishahi, Desmond Elliott

However, we do find that using an external machine translation model to generate the synthetic data sets results in better performance.

Machine Translation Representation Learning +4

Correlating neural and symbolic representations of language

1 code implementation ACL 2019 Grzegorz Chrupała, Afra Alishahi

Analysis methods which enable us to better understand the representations and functioning of neural models of language are increasingly needed as deep learning becomes the dominant approach in NLP.

Analyzing and Interpreting Neural Networks for NLP: A Report on the First BlackboxNLP Workshop

no code implementations5 Apr 2019 Afra Alishahi, Grzegorz Chrupała, Tal Linzen

The EMNLP 2018 workshop BlackboxNLP was dedicated to resources and techniques specifically developed for analyzing and understanding the inner-workings and representations acquired by neural models of language.

Symbolic inductive bias for visually grounded learning of spoken language

1 code implementation21 Dec 2018 Grzegorz Chrupała

A widespread approach to processing spoken language is to first automatically transcribe it into text.

Image Retrieval Inductive Bias +1

Revisiting the Hierarchical Multiscale LSTM

no code implementations COLING 2018 Ákos Kádár, Marc-Alexandre Côté, Grzegorz Chrupała, Afra Alishahi

Hierarchical Multiscale LSTM (Chung et al., 2016a) is a state-of-the-art language model that learns interpretable structure from character-level input.

Language Modelling

Style Obfuscation by Invariance

1 code implementation COLING 2018 Chris Emmery, Enrique Manjavacas, Grzegorz Chrupała

The task of obfuscating writing style using sequence models has previously been investigated under the framework of obfuscation-by-transfer, where the input text is explicitly rewritten in another style.

Style Transfer

On the difficulty of a distributional semantics of spoken language

no code implementations WS 2019 Grzegorz Chrupała, Lieke Gelderloos, Ákos Kádár, Afra Alishahi

In the domain of unsupervised learning most work on speech has focused on discovering low-level constructs such as phoneme inventories or word-like units.

Encoding of phonology in a recurrent neural model of grounded speech

1 code implementation CONLL 2017 Afra Alishahi, Marie Barking, Grzegorz Chrupała

We study the representation and encoding of phonemes in a recurrent neural network model of grounded speech.

Clustering

Representations of language in a model of visually grounded speech signal

2 code implementations ACL 2017 Grzegorz Chrupała, Lieke Gelderloos, Afra Alishahi

We present a visually grounded model of speech perception which projects spoken utterances and images to a joint semantic space.

From phonemes to images: levels of representation in a recurrent neural model of visually-grounded language learning

no code implementations COLING 2016 Lieke Gelderloos, Grzegorz Chrupała

We present a model of visually-grounded language learning based on stacked gated recurrent neural networks which learns to predict visual features given an image description in the form of a sequence of phonemes.

Grounded language learning

Representation of linguistic form and function in recurrent neural networks

1 code implementation CL 2017 Ákos Kádár, Grzegorz Chrupała, Afra Alishahi

We present novel methods for analyzing the activation patterns of RNNs from a linguistic point of view and explore the types of linguistic structure they learn.

Language Modelling Sentence +1

Learning language through pictures

1 code implementation IJCNLP 2015 Grzegorz Chrupała, Ákos Kádár, Afra Alishahi

We propose Imaginet, a model of learning visually grounded representations of language from coupled textual and visual input.

Sentence Word Embeddings

Text segmentation with character-level text embeddings

no code implementations18 Sep 2013 Grzegorz Chrupała

To demonstrate the usefulness of the learned text embeddings, we use them as features in a supervised character level text segmentation and labeling task: recognizing spans of text containing programming language code.

Segmentation Text Segmentation

Cannot find the paper you are looking for? You can Submit a new open access paper.