Search Results for author: María Andrea Cruz Blandón

Found 5 papers, 3 papers with code

Simultaneous or Sequential Training? How Speech Representations Cooperate in a Multi-Task Self-Supervised Learning System

no code implementations5 Jun 2023 Khazar Khorrami, María Andrea Cruz Blandón, Tuomas Virtanen, Okko Räsänen

As a result, we find that sequential training with wav2vec 2. 0 first and VGS next provides higher performance on audio-visual retrieval compared to simultaneous optimization of both learning mechanisms.

Multi-Task Learning Representation Learning +3

BabySLM: language-acquisition-friendly benchmark of self-supervised spoken language models

1 code implementation2 Jun 2023 Marvin Lavechin, Yaya Sy, Hadrien Titeux, María Andrea Cruz Blandón, Okko Räsänen, Hervé Bredin, Emmanuel Dupoux, Alejandrina Cristia

Self-supervised techniques for learning speech representations have been shown to develop linguistic competence from exposure to speech without the need for human labels.

Benchmarking Language Acquisition

Analysing the Impact of Audio Quality on the Use of Naturalistic Long-Form Recordings for Infant-Directed Speech Research

1 code implementation3 May 2023 María Andrea Cruz Blandón, Alejandrina Cristia, Okko Räsänen

Our results show that the use of modest and high audio quality naturalistic speech data result in largely similar conclusions on IDS and ADS in terms of acoustic analyses and modelling experiments.

Language Acquisition Self-Supervised Learning

Unsupervised Discovery of Recurring Speech Patterns Using Probabilistic Adaptive Metrics

2 code implementations3 Aug 2020 Okko Räsänen, María Andrea Cruz Blandón

One potential approach to this problem is to use dynamic time warping (DTW) to find well-aligning patterns from the speech data.

Dynamic Time Warping

Analysis of Predictive Coding Models for Phonemic Representation Learning in Small Datasets

no code implementations8 Jul 2020 María Andrea Cruz Blandón, Okko Räsänen

The present study investigates the behaviour of two predictive coding models, Autoregressive Predictive Coding and Contrastive Predictive Coding, in a phoneme discrimination task (ABX task) for two languages with different dataset sizes.

Language Acquisition Representation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.