1 code implementation • 2 May 2023 • Marius-Constantin Dinu, Markus Holzleitner, Maximilian Beck, Hoan Duc Nguyen, Andrea Huber, Hamid Eghbal-zadeh, Bernhard A. Moser, Sergei Pereverzyev, Sepp Hochreiter, Werner Zellinger
Our method outperforms deep embedded validation (DEV) and importance weighted validation (IWV) on all datasets, setting a new state-of-the-art performance for solving parameter choice issues in unsupervised domain adaptation with theoretical error guarantees.
1 code implementation • 25 Nov 2022 • Khaled Koutini, Shahed Masoudian, Florian Schmid, Hamid Eghbal-zadeh, Jan Schlüter, Gerhard Widmer
Furthermore, we will show that transformers trained on Audioset can be extremely effective representation extractors for a wide range of downstream tasks.
1 code implementation • 12 Jul 2022 • Christian Steinparz, Thomas Schmied, Fabian Paischer, Marius-Constantin Dinu, Vihang Patil, Angela Bitto-Nemling, Hamid Eghbal-zadeh, Sepp Hochreiter
Therefore, exploration strategies and learning methods are required that are capable of tracking the steady domain shifts, and adapting to them.
1 code implementation • 7 Jun 2022 • Martin Gauch, Maximilian Beck, Thomas Adler, Dmytro Kotsur, Stefan Fiel, Hamid Eghbal-zadeh, Johannes Brandstetter, Johannes Kofler, Markus Holzleitner, Werner Zellinger, Daniel Klotz, Sepp Hochreiter, Sebastian Lehner
We introduce SubGD, a novel few-shot learning method which is based on the recent finding that stochastic gradient descent updates tend to live in a low-dimensional parameter subspace.
2 code implementations • 24 May 2022 • Fabian Paischer, Thomas Adler, Vihang Patil, Angela Bitto-Nemling, Markus Holzleitner, Sebastian Lehner, Hamid Eghbal-zadeh, Sepp Hochreiter
We propose to utilize a frozen Pretrained Language Transformer (PLT) for history representation and compression to improve sample efficiency.
1 code implementation • NeurIPS 2021 • Werner Zellinger, Natalia Shepeleva, Marius-Constantin Dinu, Hamid Eghbal-zadeh, Hoan Nguyen, Bernhard Nessler, Sergei Pereverzyev, Bernhard A. Moser
Our approach starts with the observation that the widely-used method of minimizing the source error, penalized by a distance measure between source and target feature representations, shares characteristics with regularized ill-posed inverse problems.
2 code implementations • 8 Nov 2021 • Kajetan Schweighofer, Andreas Radler, Marius-Constantin Dinu, Markus Hofmarcher, Vihang Patil, Angela Bitto-Nemling, Hamid Eghbal-zadeh, Sepp Hochreiter
The dataset characteristics are determined by the behavioral policy that samples this dataset.
2 code implementations • 11 Oct 2021 • Khaled Koutini, Jan Schlüter, Hamid Eghbal-zadeh, Gerhard Widmer
However, one of the main shortcomings of transformer models, compared to the well-established CNNs, is the computational complexity.
Ranked #3 on Audio Classification on FSD50K (using extra training data)
no code implementations • 19 Jul 2021 • Khaled Koutini, Hamid Eghbal-zadeh, Florian Henkel, Jan Schlüter, Gerhard Widmer
Convolutional Neural Networks (CNNs) have been dominating classification tasks in various domains, such as machine vision, machine listening, and natural language processing.
1 code implementation • 26 May 2021 • Khaled Koutini, Hamid Eghbal-zadeh, Gerhard Widmer
As state-of-the-art CNN architectures-in computer vision and other domains-tend to go deeper in terms of number of layers, their RF size increases and therefore they degrade in performance in several audio classification and tagging tasks.
no code implementations • ICLR Workshop SSL-RL 2021 • Hamid Eghbal-zadeh, Florian Henkel, Gerhard Widmer
In Contextual Reinforcement Learning (CRL), a change in the context variable can cause a change in the distribution of the states.
1 code implementation • 5 Nov 2020 • Khaled Koutini, Florian Henkel, Hamid Eghbal-zadeh, Gerhard Widmer
Deep Neural Networks are known to be very demanding in terms of computing and memory requirements.
1 code implementation • 27 Jul 2020 • Khaled Koutini, Hamid Eghbal-zadeh, Verena Haunschmid, Paul Primus, Shreyan Chowdhury, Gerhard Widmer
However, the MIR field is still dominated by the classical VGG-based CNN architecture variants, often in combination with more complex modules such as attention, and/or techniques such as pre-training on large datasets.
no code implementations • 6 Jul 2020 • Hamid Eghbal-zadeh, Khaled Koutini, Paul Primus, Verena Haunschmid, Michal Lewandowski, Werner Zellinger, Bernhard A. Moser, Gerhard Widmer
Data augmentation techniques have become standard practice in deep learning, as it has been shown to greatly improve the generalisation abilities of models.
1 code implementation • 28 Oct 2019 • Khaled Koutini, Shreyan Chowdhury, Verena Haunschmid, Hamid Eghbal-zadeh, Gerhard Widmer
We present CP-JKU submission to MediaEval 2019; a Receptive Field-(RF)-regularized and Frequency-Aware CNN approach for tagging music with emotion/mood labels.
2 code implementations • 5 Sep 2019 • Khaled Koutini, Hamid Eghbal-zadeh, Gerhard Widmer
One side effect of restricting the RF of CNNs is that more frequency information is lost.
1 code implementation • 4 Sep 2019 • Paul Primus, Hamid Eghbal-zadeh, David Eitelsebner, Khaled Koutini, Andreas Arzt, Gerhard Widmer
Distribution mismatches between the data seen at training and at application time remain a major challenge in all application areas of machine learning.
3 code implementations • 3 Jul 2019 • Khaled Koutini, Hamid Eghbal-zadeh, Matthias Dorfer, Gerhard Widmer
To this end, we analyse the receptive field (RF) of these CNNs and demonstrate the importance of the RF to the generalization capability of the models.
no code implementations • 16 May 2019 • Hamid Eghbal-zadeh, Lukas Fischer, Thomas Hoch
Additionally, we show that the O-GAN achieves better conditioning results evaluated by implicit similarity between the text and the generated image.
1 code implementation • CVPR 2019 • Hamid Eghbal-zadeh, Werner Zellinger, Gerhard Widmer
Generative Adversarial Networks have surprising ability for generating sharp and realistic images, though they are known to suffer from the so-called mode collapse problem.
1 code implementation • 22 Jun 2018 • Hamid Eghbal-zadeh, Lukas Fischer, Niko Popitsch, Florian Kromp, Sabine Taschner-Mandl, Khaled Koutini, Teresa Gerber, Eva Bozsaky, Peter F. Ambros, Inge M. Ambros, Gerhard Widmer, Bernhard A. Moser
We show, that Deep SNP is capable of successfully predicting the presence or absence of a breakpoint in large genomic windows and outperforms state-of-the-art neural network models.
no code implementations • 10 Nov 2017 • Hamid Eghbal-zadeh, Matthias Dorfer, Gerhard Widmer
To tackle this problem, we propose Deep Within-Class Covariance Analysis (DWCCA), a deep neural network layer that significantly reduces the within-class covariance of a DNN's representation, improving performance on unseen test data from a shifted distribution.
no code implementations • 6 Aug 2017 • Hamid Eghbal-zadeh, Gerhard Widmer
The central idea is to integrate a probabilistic model (a Gaussian Mixture Model, in our case) into the GAN framework which supports a new kind of loss function (based on likelihood rather than classification loss), and at the same time gives a meaningful measure of the quality of the outputs generated by the network.
no code implementations • 24 Jul 2017 • Hamid Eghbal-zadeh, Gerhard Widmer
We present a simple method for assessing the quality of generated images in Generative Adversarial Networks (GANs).
no code implementations • 20 Jun 2017 • Hamid Eghbal-zadeh, Bernhard Lehner, Matthias Dorfer, Gerhard Widmer
Finally, we propose a hybrid system for ASC using multi-channel i-vectors and CNNs by utilizing a score fusion technique.