Search Results for author: Shreyan Chowdhury

Found 9 papers, 4 papers with code

Expressivity-aware Music Performance Retrieval using Mid-level Perceptual Features and Emotion Word Embeddings

no code implementations26 Jan 2024 Shreyan Chowdhury, Gerhard Widmer

On the text side, we use emotion-enriched word embeddings (EWE) and on the audio side, we extract mid-level perceptual features instead of generic audio embeddings.

Retrieval Word Embeddings

Are we describing the same sound? An analysis of word embedding spaces of expressive piano performance

1 code implementation31 Dec 2023 Silvan David Peter, Shreyan Chowdhury, Carlos Eduardo Cancino-Chacón, Gerhard Widmer

Using a music research dataset of free text performance characterizations and a follow-up study sorting the annotations into clusters, we derive a ground truth for a domain-specific semantic similarity structure.

Information Retrieval Retrieval +2

Tracing Back Music Emotion Predictions to Sound Sources and Intuitive Perceptual Qualities

2 code implementations14 Jun 2021 Shreyan Chowdhury, Verena Praher, Gerhard Widmer

In previous work, we have shown how to derive explanations of model predictions in terms of spectrogram image segments that connect to the high-level emotion prediction via a layer of easily interpretable perceptual features.

Emotion Recognition Information Retrieval +3

On the Characterization of Expressive Performance in Classical Music: First Results of the Con Espressione Game

no code implementations5 Aug 2020 Carlos Cancino-Chacón, Silvan Peter, Shreyan Chowdhury, Anna Aljanaki, Gerhard Widmer

In this paper, we offer a first account of this new data resource for expressive performance research, and provide an exploratory analysis, addressing three main questions: (1) how similarly do different listeners describe a performance of a piece?

Descriptive

Receptive-Field Regularized CNNs for Music Classification and Tagging

1 code implementation27 Jul 2020 Khaled Koutini, Hamid Eghbal-zadeh, Verena Haunschmid, Paul Primus, Shreyan Chowdhury, Gerhard Widmer

However, the MIR field is still dominated by the classical VGG-based CNN architecture variants, often in combination with more complex modules such as attention, and/or techniques such as pre-training on large datasets.

Classification General Classification +4

Emotion and Theme Recognition in Music with Frequency-Aware RF-Regularized CNNs

1 code implementation28 Oct 2019 Khaled Koutini, Shreyan Chowdhury, Verena Haunschmid, Hamid Eghbal-zadeh, Gerhard Widmer

We present CP-JKU submission to MediaEval 2019; a Receptive Field-(RF)-regularized and Frequency-Aware CNN approach for tagging music with emotion/mood labels.

Acoustic Scene Classification Scene Classification

Two-level Explanations in Music Emotion Recognition

no code implementations28 May 2019 Verena Haunschmid, Shreyan Chowdhury, Gerhard Widmer

Current ML models for music emotion recognition, while generally working quite well, do not give meaningful or intuitive explanations for their predictions.

Emotion Recognition Music Emotion Recognition +1

Cannot find the paper you are looking for? You can Submit a new open access paper.