Search Results for author: Keizo Oyama

Found 6 papers, 1 papers with code

Interpretable Melody Generation from Lyrics with Discrete-Valued Adversarial Training

no code implementations30 Jun 2022 Wei Duan, Zhe Zhang, Yi Yu, Keizo Oyama

Generating melody from lyrics is an interesting yet challenging task in the area of artificial intelligence and music.

MusicTM-Dataset for Joint Representation Learning among Sheet Music, Lyrics, and Musical Audio

no code implementations1 Dec 2020 Donghuo Zeng, Yi Yu, Keizo Oyama

This work present a music dataset named MusicTM-Dataset, which is utilized in improving the representation learning ability of different types of cross-modal retrieval (CMR).

Cross-Modal Retrieval Information Retrieval +3

Unsupervised Generative Adversarial Alignment Representation for Sheet music, Audio and Lyrics

no code implementations29 Jul 2020 Donghuo Zeng, Yi Yu, Keizo Oyama

In this paper, we propose an unsupervised generative adversarial alignment representation (UGAAR) model to learn deep discriminative representations shared across three major musical modalities: sheet music, lyrics, and audio, where a deep neural network based architecture on three branches is jointly trained.

Representation Learning

Audio-Visual Embedding for Cross-Modal MusicVideo Retrieval through Supervised Deep CCA

no code implementations10 Aug 2019 Donghuo Zeng, Yi Yu, Keizo Oyama

ii) We propose an end-to-end deep model for cross-modal audio-visual learning where S-DCCA is trained to learn the semantic correlation between audio and visual modalities.

audio-visual learning Retrieval +1

Personalized Music Recommendation with Triplet Network

no code implementations10 Aug 2019 Haoting Liang, Donghuo Zeng, Yi Yu, Keizo Oyama

Since many online music services emerged in recent years so that effective music recommendation systems are desirable.

Music Recommendation Recommendation Systems

Deep Triplet Neural Networks with Cluster-CCA for Audio-Visual Cross-modal Retrieval

2 code implementations10 Aug 2019 Donghuo Zeng, Yi Yu, Keizo Oyama

In particular, two significant contributions are made: i) a better representation by constructing deep triplet neural network with triplet loss for optimal projections can be generated to maximize correlation in the shared subspace.

Cross-Modal Retrieval Information Retrieval +1

Cannot find the paper you are looking for? You can Submit a new open access paper.