Search Results for author: Seungheon Doh

Found 8 papers, 5 papers with code

The Song Describer Dataset: a Corpus of Audio Captions for Music-and-Language Evaluation

1 code implementation16 Nov 2023 Ilaria Manco, Benno Weck, Seungheon Doh, Minz Won, Yixiao Zhang, Dmitry Bogdanov, Yusong Wu, Ke Chen, Philip Tovstogan, Emmanouil Benetos, Elio Quinton, György Fazekas, Juhan Nam

We introduce the Song Describer dataset (SDD), a new crowdsourced corpus of high-quality audio-caption pairs, designed for the evaluation of music-and-language models.

Music Captioning Music Generation +2

LP-MusicCaps: LLM-Based Pseudo Music Captioning

1 code implementation31 Jul 2023 Seungheon Doh, Keunwoo Choi, Jongpil Lee, Juhan Nam

In addition, we trained a transformer-based music captioning model with the dataset and evaluated it under zero-shot and transfer-learning settings.

Language Modelling Large Language Model +3

Textless Speech-to-Music Retrieval Using Emotion Similarity

no code implementations19 Mar 2023 Seungheon Doh, Minz Won, Keunwoo Choi, Juhan Nam

We introduce a framework that recommends music based on the emotions of speech.

Retrieval

Music Playlist Title Generation Using Artist Information

1 code implementation14 Jan 2023 Haven Kim, Seungheon Doh, Junwon Lee, Juhan Nam

Automatically generating or captioning music playlist titles given a set of tracks is of significant interest in music streaming services as customized playlists are widely used in personalized music recommendation, and well-composed text titles attract users and help their music discovery.

Music Recommendation

Toward Universal Text-to-Music Retrieval

3 code implementations26 Nov 2022 Seungheon Doh, Minz Won, Keunwoo Choi, Juhan Nam

This paper introduces effective design choices for text-to-music retrieval systems.

Music Classification Retrieval +2

TräumerAI: Dreaming Music with StyleGAN

1 code implementation9 Feb 2021 Dasaem Jeong, Seungheon Doh, Taegyun Kwon

The goal of this paper to generate a visually appealing video that responds to music with a neural network so that each frame of the video reflects the musical characteristics of the corresponding audio clip.

 Ranked #1 on Music Auto-Tagging on TimeTravel (using extra training data)

Music Auto-Tagging

Musical Word Embedding: Bridging the Gap between Listening Contexts and Music

no code implementations23 Jul 2020 Seungheon Doh, Jongpil Lee, Tae Hong Park, Juhan Nam

Word embedding pioneered by Mikolov et al. is a staple technique for word representations in natural language processing (NLP) research which has also found popularity in music information retrieval tasks.

Information Retrieval Music Information Retrieval +1

Cannot find the paper you are looking for? You can Submit a new open access paper.