Music Generation
129 papers with code • 0 benchmarks • 24 datasets
Music Generation is the task of generating music or music-like sounds from a model or algorithm. The goal is to produce a sequence of notes or sound events that are similar to existing music in some way, such as having the same style, genre, or mood.
Benchmarks
These leaderboards are used to track progress in Music Generation
Libraries
Use these libraries to find Music Generation models and implementationsDatasets
Latest papers with no code
Long-form music generation with latent diffusion
Audio-based generative models for music have seen great strides recently, but so far have not managed to produce full-length music tracks with coherent musical structure.
Text-to-Song: Towards Controllable Music Generation Incorporating Vocals and Accompaniment
A song is a combination of singing voice and accompaniment.
MuPT: A Generative Symbolic Music Pretrained Transformer
In this paper, we explore the application of Large Language Models (LLMs) to the pre-training of music.
A Novel Bi-LSTM And Transformer Architecture For Generating Tabla Music
In this technical paper, methods for generating classical Indian music, specifically tabla music, is proposed.
The NES Video-Music Database: A Dataset of Symbolic Video Game Music Paired with Gameplay Videos
To address this research gap, we introduce a novel dataset named NES-VMDB, containing 98, 940 gameplay videos from 389 NES games, each paired with its original soundtrack in symbolic format (MIDI).
Motifs, Phrases, and Beyond: The Modelling of Structure in Symbolic Music Generation
Modelling musical structure is vital yet challenging for artificial intelligence systems that generate symbolic music compositions.
ByteComposer: a Human-like Melody Composition Method based on Language Model Agent
Large Language Models (LLM) have shown encouraging progress in multimodal understanding and generation tasks.
A Survey of Music Generation in the Context of Interaction
In recent years, machine learning, and in particular generative adversarial neural networks (GANs) and attention-based neural networks (transformers), have been successfully used to compose and generate music, both melodies and polyphonic pieces.
Structure-informed Positional Encoding for Music Generation
Music generated by deep learning methods often suffers from a lack of coherence and long-term organization.
An Order-Complexity Aesthetic Assessment Model for Aesthetic-aware Music Recommendation
In order to improve the quality of AI music generation and further guide computer music production, synthesis, recommendation and other tasks, we use Birkhoff's aesthetic measure to design a aesthetic model, objectively measuring the aesthetic beauty of music, and form a recommendation list according to the aesthetic feeling of music.