Audio Signal Processing
20 papers with code • 0 benchmarks • 2 datasets
This is a general task that covers transforming audio inputs into audio outputs, not limited to existing PaperWithCode categories of Source Separation, Denoising, Classification, Recognition, etc.
Benchmarks
These leaderboards are used to track progress in Audio Signal Processing
Most implemented papers
L3DAS21 Challenge: Machine Learning for 3D Audio Signal Processing
The L3DAS21 Challenge is aimed at encouraging and fostering collaborative research on machine learning for 3D audio signal processing, with particular focus on 3D speech enhancement (SE) and 3D sound localization and detection (SELD).
DeepSpectrumLite: A Power-Efficient Transfer Learning Framework for Embedded Speech and Audio Processing from Decentralised Data
By obtaining state-of-the-art results on a set of paralinguistics tasks, we demonstrate the suitability of the proposed transfer learning approach for embedded audio signal processing, even when data is scarce.
Visualization of Linear Operations in the Spherical Harmonics Domain
Linear operations on coefficients in the spherical harmonics (SH) transform domain that again yield SH-domain coefficients are an important toolset in many disciplines of research and engineering.
Sound2Synth: Interpreting Sound via FM Synthesizer Parameters Estimation
Synthesizer is a type of electronic musical instrument that is now widely used in modern music production and sound design.
MYRiAD: A Multi-Array Room Acoustic Database
In the development of acoustic signal processing algorithms, their evaluation in various acoustic environments is of utmost importance.
Subspace-Configurable Networks
The obtained subspace is low-dimensional and has a surprisingly simple structure even for complex, non-invertible transformations of the input, leading to an exceptionally high efficiency of subspace-configurable networks (SCNs) when limited storage and computing resources are at stake.
MF-PAM: Accurate Pitch Estimation through Periodicity Analysis and Multi-level Feature Fusion
We introduce Multi-level feature Fusion-based Periodicity Analysis Model (MF-PAM), a novel deep learning-based pitch estimation model that accurately estimates pitch trajectory in noisy and reverberant acoustic environments.
Instabilities in Convnets for Raw Audio
Numerical simulations align with our theory and suggest that the condition number of a convolutional layer follows a logarithmic scaling law between the number and length of the filters, which is reminiscent of discrete wavelet bases.
Unsupervised Harmonic Parameter Estimation Using Differentiable DSP and Spectral Optimal Transport
In neural audio signal processing, pitch conditioning has been used to enhance the performance of synthesizers.
A Survey on Data Augmentation in Large Model Era
Leveraging large models, these data augmentation techniques have outperformed traditional approaches.