Clara: A Neural Net Music Generator
Take the AI vs Human Quiz.
- PyTorch version 0.3.0.post4
- GeneralUser: install in data folder (this is needed to translate midi files to mp3)
- MIT's music21
ln -s ./replace/this/with/your/path/to/fastai/library fastai
to create a symbolic link to the fastai library. Alternately, this blog has a clear description of how to get an AWS machine up and running with FastAI already good to go.
You will also likely need to use sudo apt install to get fluidsynth, mpg321, and twolame.
Run the Jupyter Notebook BasicIntro.ipynb or follow the individual instructions here. To create generations with a pretrained notewise model, using only the default settings, run: Basic:
python make_test_train.py --example python generate.py -model notewise_generator -output notewise_generation_samples
The output samples will be in data/output/notewise_generation_samples, or open Playlist.ipynb to listen to the output samples. I recommend the free program MuseScore to translate the midi files into sheet music.
Note, you must first make sure the requirements (above) are installed.
If you use your own midi files, they should go in data/composers/midi/piano_solo or data/composers/midi/chamber (the project expects to see a folder of midi files for each composer, ie: data/composers/midi/piano_solo/bach/example_piece.mid). Data:
to translate midi files to text files in the various notewise and chordwise options.
My dataset is available here (you can download any or all):
Put these in data/composers/notewise:
- Chordwise piano solo text files
- Chordwise jazz piano solo text files
- Chordwise piano/violin text files
Training and Generation:
- make_test_train.py - create the training and testing datasets (adjust notewise/chordwise, optionally create only a small sample size)
- train.py - train an AWD-LSTM (adjust model parameters, dropout, and training regime)
- generate.py - generate new samples (adjust generation size)
If you use the data files I've linked above, those are quite large, and will take a long time to train. If you are looking to experiment with different training networks, I'd highly recommend at first using --sample .1 (10% of the data) with make_test_train.py, so that you have a much smaller dataset to play with and can iterate faster.
Playlist.ipynb is a simple Jupyter Notebook which creates a nicely formatted playlist for listening to all the generations.
- make_critic_data.py - create the training and test datasests (requires a trained generation model to create the fake data)
- critic.py - trains a classifier to predict if a sample is human-composed or LSTM-composed
- make_composer_data.py - create the training and test datasests (all from human composed pieces)
- composer_classifier.py - trains a classifier to predict which human composed the piece
Sample pretrained models are included in this repository. They were trained using the default settings (all composers, notewise using a sample frequency 12, chordwise using a sample frequency 4). Pretrained Models:
- chamber_generator (uses notewise encoding)
For example, use:
python generator.py -model notewise_generator -output notewise_generation_samples --random_freq 0.8 --trunc 3
to generate musical samples.