Search Results for author: Anderson R. Avila

Found 8 papers, 1 papers with code

On the Impact of Quantization and Pruning of Self-Supervised Speech Models for Downstream Speech Recognition Tasks "In-the-Wild''

no code implementations25 Sep 2023 Arthur Pimentel, Heitor Guimarães, Anderson R. Avila, Mehdi Rezagholizadeh, Tiago H. Falk

Recent advances with self-supervised learning have allowed speech recognition systems to achieve state-of-the-art (SOTA) word error rates (WER) while requiring only a fraction of the labeled training data needed by its predecessors.

Data Augmentation Model Compression +4

Multimodal Audio-textual Architecture for Robust Spoken Language Understanding

no code implementations12 Jun 2023 Anderson R. Avila, Mehdi Rezagholizadeh, Chao Xing

In this work, we investigate impacts of this ASR error propagation on state-of-the-art NLU systems based on pre-trained language models (PLM), such as BERT and RoBERTa.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +3

RobustDistiller: Compressing Universal Speech Representations for Enhanced Environment Robustness

no code implementations18 Feb 2023 Heitor R. Guimarães, Arthur Pimentel, Anderson R. Avila, Mehdi Rezagholizadeh, Boxing Chen, Tiago H. Falk

The proposed layer-wise distillation recipe is evaluated on top of three well-established universal representations, as well as with three downstream tasks.

Knowledge Distillation Multi-Task Learning

Improving the Robustness of DistilHuBERT to Unseen Noisy Conditions via Data Augmentation, Curriculum Learning, and Multi-Task Enhancement

no code implementations12 Nov 2022 Heitor R. Guimarães, Arthur Pimentel, Anderson R. Avila, Mehdi Rezagholizadeh, Tiago H. Falk

Self-supervised speech representation learning aims to extract meaningful factors from the speech signal that can later be used across different downstream tasks, such as speech and/or emotion recognition.

Data Augmentation Emotion Recognition +2

Sequential End-to-End Intent and Slot Label Classification and Localization

no code implementations8 Jun 2021 Yiran Cao, Nihal Potdar, Anderson R. Avila

Such approaches allow for the extraction of semantic information directly from the speech signal, thus bypassing the need for a transcript from an automatic speech recognition (ASR) system.

Ranked #11 on Spoken Language Understanding on Fluent Speech Commands (using extra training data)

Automatic Speech Recognition Automatic Speech Recognition (ASR) +4

A Streaming End-to-End Framework For Spoken Language Understanding

no code implementations20 May 2021 Nihal Potdar, Anderson R. Avila, Chao Xing, Dong Wang, Yiran Cao, Xiao Chen

In this paper, we propose a streaming end-to-end framework that can process multiple intentions in an online and incremental way.

Intent Detection Keyword Spotting +3

Cannot find the paper you are looking for? You can Submit a new open access paper.