1 code implementation • 28 Oct 2022 • Louis Bahrman, Marina Krémé, Paul Magron, Antoine Deleforge
Signal inpainting is the task of restoring degraded or missing samples in a signal.
no code implementations • 1 Sep 2021 • Cédric Foy, Antoine Deleforge, Diego Di Carlo
The critical choices of geometric, acoustic and simulation parameters used to train the models are extensively discussed and studied, while keeping in mind conditions that are representative of the field of building acoustics.
1 code implementation • 29 Jul 2021 • Prerak Srivastava, Antoine Deleforge, Emmanuel Vincent
Knowing the geometrical and acoustical parameters of a room may benefit applications such as audio augmented reality, speech dereverberation or audio forensics.
1 code implementation • 13 Jun 2021 • Archontis Politis, Sharath Adavanne, Daniel Krause, Antoine Deleforge, Prerak Srivastava, Tuomas Virtanen
This report presents the dataset and baseline of Task 3 of the DCASE2021 Challenge on Sound Event Localization and Detection (SELD).
2 code implementations • 27 Apr 2021 • Diego Di Carlo, Pinchas Tandeitnik, Cédric Foy, Antoine Deleforge, Nancy Bertin, Sharon Gannot
This paper presents dEchorate: a new database of measured multichannel Room Impulse Responses (RIRs) including annotations of early echo timings and 3D positions of microphones, real sources and image sources under different wall configurations in a cuboid room.
5 code implementations • 22 May 2020 • Joris Cosentino, Manuel Pariente, Samuele Cornell, Antoine Deleforge, Emmanuel Vincent
Most deep learning-based speech separation models today are benchmarked on it.
Audio and Speech Processing
2 code implementations • 23 Oct 2019 • Manuel Pariente, Samuele Cornell, Antoine Deleforge, Emmanuel Vincent
Also, we validate the use of parameterized filterbanks and show that complex-valued representations and masks are beneficial in all conditions.
no code implementations • 3 May 2019 • Manuel Pariente, Antoine Deleforge, Emmanuel Vincent
Recent studies have explored the use of deep generative models of speech spectra based of variational autoencoders (VAEs), combined with unsupervised noise models, to perform speech enhancement.
1 code implementation • NeurIPS 2018 • Helena Peic Tukuljac, Antoine Deleforge, Rémi Gribonval
The approach operates directly in the parameter-space of echo locations and weights, and enables near-exact blind and off-grid echo retrieval from discrete-time measurements.
1 code implementation • 13 Nov 2017 • Nicolas Keriven, Antoine Deleforge, Antoine Liutkus
We propose a new blind source separation algorithm based on mixtures of alpha-stable distributions.
no code implementations • 14 Dec 2016 • Clément Gaultier, Saurabh Kataria, Antoine Deleforge
This paper introduces a new paradigm for sound source lo-calization referred to as virtual acoustic space traveling (VAST) and presents a first dataset designed for this purpose.
no code implementations • 30 Sep 2016 • Antoine Deleforge, Yann Traonmilin
We consider the problem of estimating the phases of K mixed complex signals from a multichannel observation, when the mixing matrix and signal magnitudes are known.
no code implementations • 31 Mar 2016 • Vincent Drouard, Radu Horaud, Antoine Deleforge, Silèye Ba, Georgios Evangelidis
Head-pose estimation has many applications, such as social event analysis, human-robot and human-computer interaction, driving assistance, and so forth.
no code implementations • 30 Sep 2014 • Antoine Deleforge, Florence Forbes, Sileye . Ba, Radu Horaud
This involves resolving inverse problems which can be addressed within machine learning, with the advantage that, once a relationship between physical parameters and spectra has been established in a data-driven fashion, the learned relationship can be used to estimate physical parameters for new hyper-spectral observations.
no code implementations • 12 Aug 2014 • Antoine Deleforge, Radu Horaud, Yoav Schechner, Laurent Girin
Indeed, we demonstrate that the method can be used for audio-visual fusion, namely to map speech signals onto images and hence to spatially align the audio and visual modalities, thus enabling to discriminate between speaking and non-speaking faces.
no code implementations • 10 Aug 2013 • Antoine Deleforge, Florence Forbes, Radu Horaud
We introduce a mixture of locally-linear probabilistic mapping model that starts with estimating the parameters of inverse regression, and follows with inferring closed-form solutions for the forward parameters of the high-dimensional regression problem of interest.