Skip to content

bgenchel/Musical-SeqGAN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

88 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Curious-Musical-SeqGAN

Adapt and evaluate SeqGAN (https://arxiv.org/abs/1609.05473) for music generation.

The original paper briefly mentions that the model was trained for music generation on the Nottingham folk music dataset, and evaluated using BLEU score. Unfortunately, there was not much information given on how the model was adapted for this use case, and how they acheived the scores they reported. Additionally, code from the authors and others are implemented exclusively towards the central task used in the paper, learning to model a randomly initialized LSTM.

Here, we attempt to adapt the model specifically for this purpose and give a clearer and more detailed evaluation on the task of music generation.

N.B. When cloning this repo, use the flag --recursive after the address, i.e. git clone https://github.com/bgenchel/MusicalSeqGAN.git --recursive.

Otherwise, run git submodule update --init in the project root. This is to fetch the git submodule for MGEval.

Running SeqGAN on Nottingham

First, create a conda environment based off of the included environment.yml file by running:

conda env create -f environment.yml

The rest of the commands should be done with this environment active.

To fetch the datasets, run the following:

  • From project root, cd src/scripts
  • Then run python get_data.py

To train the model, run the following:

  • From project root, cd src/models/nottingham
  • Then run python main.py

Running the MGEval Toolkit

The MGEval toolkit requires the use of python 2, which is not the version of python needed to run SeqGAN.

The following steps outline how to create an environment suitable to run the toolkit.

N.B. This environment does not work on Mac OSX due to matplotlib needing python as a framework.

Please do this on a Linux machine if possible.

Replace ENV_NAME with a name you see fit.

  • conda create --name ENV_NAME python=2
  • conda install scikit-learn
  • pip install matplotlib
  • pip install seaborn
  • pip install pretty_midi
  • pip install python-midi
  • pip install metrics

To generate MIDI files first run the following using the python 3 rl4gm env:

  • From project root, cd src/models/nottingham
  • (rl4gm) python eval.py -r
    • This defaults to generating 1000 sequences each for targets, pretrained generator, and adversarially trained generator.
    • Use --num_midi_samples NUM to change the number of MIDI files generated.

Then, using the python 2 env, run the following:

  • From project root, cd src/evaluation
  • (python2) python toolkit.py
    • The MGEval toolkit is not consistent on what it requires to generate/display each metric.
    • It is currently set up to show the note_length_transition_matrix for the targets and adversarially trained generator samples.
    • To explore further, change the metric name, and update the expected shape and args/kwargs as needed, per the source here.

About

Adapt and evaluate SeqGAN for music. Developed in PyTorch, using https://github.com/ZiJianZhao/SeqGAN-PyTorch as a base

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published