Speaker Separation

11 papers with code • 0 benchmarks • 3 datasets

This task has no description! Would you like to contribute one?

Latest papers with no code

New Insights on Target Speaker Extraction

no code yet • 1 Feb 2022

However, such studies have been conducted on a few datasets and have not considered recent deep neural network architectures for SS that have shown impressive separation performance.

Location-based training for multi-channel talker-independent speaker separation

no code yet • 8 Oct 2021

We further demonstrate the effectiveness of LBT for the separation of four and five concurrent speakers.

Online Self-Attentive Gated RNNs for Real-Time Speaker Separation

no code yet • 25 Jun 2021

Our stateful implementation for online separation leads to a minor drop in performance compared to the offline model; 0. 8dB for monaural inputs and 0. 3dB for binaural inputs while reaching a real-time factor of 0. 65.

Personalized Keyphrase Detection using Speaker and Environment Information

no code yet • 28 Apr 2021

In this paper, we introduce a streaming keyphrase detection system that can be easily customized to accurately detect any phrase composed of words from a large vocabulary.

Guided Training: A Simple Method for Single-channel Speaker Separation

no code yet • 26 Mar 2021

Another way is to use an anchor speech, a short speech of the target speaker, to model the speaker identity.

On permutation invariant training for speech source separation

no code yet • 9 Feb 2021

We study permutation invariant training (PIT), which targets at the permutation ambiguity problem for speaker independent source separation models.

High Fidelity Speech Regeneration with Application to Speech Enhancement

no code yet • 31 Jan 2021

Speech enhancement has seen great improvement in recent years mainly through contributions in denoising, speaker separation, and dereverberation methods that mostly deal with environmental effects on vocal audio.

Interactive Speech and Noise Modeling for Speech Enhancement

no code yet • 17 Dec 2020

In this paper, we propose a novel idea to model speech and noise simultaneously in a two-branch convolutional neural network, namely SN-Net.

Informed Source Extraction With Application to Acoustic Echo Reduction

no code yet • 9 Nov 2020

Recent deep learning-based methods leverage a speaker discriminative model that maps a reference snippet uttered by the target speaker into a single embedding vector that encapsulates the characteristics of the target speaker.

Speaker Separation Using Speaker Inventories and Estimated Speech

no code yet • 20 Oct 2020

We propose speaker separation using speaker inventories and estimated speech (SSUSIES), a framework leveraging speaker profiles and estimated speech for speaker separation.