AI-generated music challenge
HiddenMusic generated by AI
Can AI generate music that we humans find beautiful, perhaps even moving? Let’s find out!
In this challenge, participants are tasked to generate an AI model that learns on a large data set of music (in the form of MIDI files), and is then capable of producing its own music. Concretely, the model must produce a music piece in response to a short “seed” MIDI file that is given as input.
There are two special aspects of this challenge, apart from the extremely interesting application. First, the results of the models will be evaluated by humans, with an ELO-style system where volunteers are given two randomly paired pieces of generated music, and choose the one they like better. Second, the top five models will at the end each generate a piece of music that will be performed live on stage at the Applied Machine Learning Days!
Evaluation criteria
The grader expects a MIDI file of a total length of 3600
seconds (when played at 120 bpm
). The MIDI file has to be a type 0
MIDI file (a maximum of 1 track), and in the case of multiple tracks, only the first track will be considered. There are no challenge-specific restrictions on the number of channels being used in the MIDI file.
The grader splits the MIDI file into 120 chunks
of approximately 30 seconds
each, and each submission is represented by this pool of 120 chunks.
During this post-processing step, all meta events from the MIDI file will be removed except the PPQ meta event (or ticks per beat
), hence the officially supported MIDI events will only be the note_on
and note_off
events; where note_off
event can be optionally replaced by a note_on
event with a velocity of 0
. All the MIDI parsing is done using the MIDO library; and you are requested to ensure that your submitted file is estimated to be of 3600 +/- 10 seconds
by mido.MidiFile('your_file_path').length
.
A separate evaluation interface is made available, where all the participants (and other external volunteers) can hear two randomly sampled chunks and then vote for the one they like better (more details on the sampling mechanism is provided in the following sections). These randomly sampled chunks will be played with the SoundFont of an acoustic grand piano at 120 bpm
.
These binary comparisons will be used to compute an individual score for every submission, which evolves over time as it gets more and more evaluations in the evaluation interface. The scoring mechanism follows the TrueSkill ranking system, and hence is modeled by
The submissions tab will report the values for
To ensure that the top-10 selected participants are not overfitting on the training set; the top-10 submissions at the end of the challenge, will be divided into quantized chunks of
where,
All matching chunks pairs with
Starter Kit : A starter kit to help you get started on the submission procedure is made available at: https://github.com/crowdAI/crowdai-ai-generate-music-starter-kit.
Comin Soon : A Getting Started guide on music generation from MIDI files using LSTMs.
Resources
Starter Kit : A starter kit to help you get started on the submission procedure is made available at : https://github.com/crowdAI/crowdai-ai-generate-music-starter-kit.
Some other projects to help you quickly get started on MIDI composition:
- https://github.com/brannondorsey/midi-rnn
- https://github.com/jisungk/deepjazz
- Google Magenta : Performance RNN
- MIDINet
Contact:
- Technical issues : https://gitter.im/crowdAI/AI-Generated-Music-Challenge
- Discussion Forum : https://www.crowdai.org/challenges/ai-generated-music-challenge/topics
We strongly encourage you to use the public channels mentioned above for communications between the participants and the organizers. In extreme cases, if there are any queries or comments that you would like to make using a private communication channel, then you can send us an email at :
- Sharada Prasanna Mohanty sharada.mohanty@epfl.ch
- Florian Colombo florian.colombo@epfl.ch
Prizes
Datasets License
abcd
Participants















