aiexperiments-ai-duet/server/third_party/magenta/models/basic_rnn
2016-11-11 15:34:34 -05:00
..
__init__.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
basic_rnn_create_dataset.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
basic_rnn_encoder_decoder.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
basic_rnn_generate.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
basic_rnn_generator.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
basic_rnn_graph.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
basic_rnn_train.py moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
BUILD moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
README.md moving magenta to third_party folder 2016-11-11 15:34:34 -05:00
run_basic_rnn_train.sh moving magenta to third_party folder 2016-11-11 15:34:34 -05:00

Basic RNN

This model provides baselines for the application of language modeling to melody generation. This code also serves as a working example for implementing a language model in TensorFlow. In this code, an LSTM cell is used, but any type of cell can be swapped in.

This model takes monophonic melodies, meaning one note plays at a time. Use basic_rnn_create_dataset.py to extract monophonic melodies from NoteSequence protos made from your MIDI files. The script will look at each instrument track and extract a melody line if it is at least 7 measures long, and at least 5 unique pitches (with octave equivalence). If multiple notes play at the same time, one note is kept.

How to Use

First, set up your Magenta environment. Next, you can either use a pre-trained model or train your own.

Pre-trained

If you want to get started right away, you can use a model that we've pre-trained on thousands of MIDI files. Download the basic_rnn bundle.

Generate a melody

BUNDLE_PATH=<absolute path of basic_rnn.mag>

bazel run //magenta/models/basic_rnn:basic_rnn_generate -- \
--bundle_file=${BUNDLE_PATH} \
--output_dir=/tmp/basic_rnn/generated \
--num_outputs=10 \
--num_steps=128 \
--primer_melody="[60]"

This will generate a melody starting with a middle C. If you'd like, you can also supply a priming melody using a string representation of a Python list. The values in the list should be ints that follow the melodies_lib.Melody format (-2 = no event, -1 = note-off event, values 0 through 127 = note-on event for that MIDI pitch). For example --primer_melody="[60, -2, 60, -2, 67, -2, 67, -2]" would prime the model with the first four notes of Twinkle Twinkle Little Star. Instead of using --primer_melody, we can use --primer_midi to prime our model with a melody stored in a MIDI file. For example, --primer_midi=<absolute path to magenta/models/shared/primer.mid> will prime the model with the melody in that MIDI file.

Train your own

First, you'll need to create a training datset. Next, you'll train your model on this data.

Run basic_rnn_create_dataset.py on the sequences dataset that is generated by convert_midi_dir_to_note_sequences.py as shown below. This will extract melodies from NoteSequence data (which was extracted from MIDI data). The output is written to disk as a tfrecord file that contains SequenceExample protos. TensorFlow readers in the basic_rnn model can read SequenceExample protos from disk directly into the model. In this example, we create an evaluation dataset in a second tfrecord file, but that can be omitted by leaving out the eval_output and eval_ratio flags.

# TFRecord file containing NoteSequence protocol buffers from convert_midi_dir_to_note_sequences.py.
SEQUENCES_TFRECORD=/tmp/notesequences.tfrecord

# Where training and evaluation datasets will be written.
DATASET_DIR=/tmp/basic_rnn/sequence_examples

# TFRecord file that TensorFlow's SequenceExample protos will be written to. This is the training dataset.
TRAIN_DATA=$DATASET_DIR/training_melodies.tfrecord

# Optional evaluation dataset. Also, a TFRecord file containing SequenceExample protos.
EVAL_DATA=$DATASET_DIR/eval_melodies.tfrecord

# Fraction of input data that will be written to the eval dataset (if eval_output flag is set).
EVAL_RATIO=0.10

bazel run //magenta/models/basic_rnn:basic_rnn_create_dataset -- \
--input=$SEQUENCES_TFRECORD \
--output_dir=$DATASET_DIR \
--eval_ratio=$EVAL_RATIO

Running training in depth

Build basic_rnn_train first so that it can be run multiple times in parallel.

bazel build //magenta/models/basic_rnn:basic_rnn_train

Save train and eval datasets as /tmp/basic_rnn/sequence_examples/training_melodies.tfrecord and /tmp/basic_rnn/sequence_examples/eval_melodies.tfrecord.

Create an experiment directory, say /tmp/basic_rnn/logdir, and choose a subdirectory to save this run in, like /tmp/basic_rnn/logdir/run1. Increase the number to run2, run3, etc... every time you rerun the same experiment, so that you don't clobber previous experiment output.

Lets create an LSTM model with 1 cell of size 50. So the hyperparameter string is '{"rnn_layer_sizes":[50]}'.

Run training job from the project root

./bazel-bin/magenta/models/basic_rnn/basic_rnn_train --run_dir=/tmp/basic_rnn/logdir/run1 --sequence_example_file=$TRAIN_DATA --hparams='{"rnn_layer_sizes":[50]}' --num_training_steps=20000

Optionally run eval job in parallel with training job

./bazel-bin/magenta/models/basic_rnn/basic_rnn_train --run_dir=/tmp/basic_rnn/logdir/run1 --sequence_example_file=$EVAL_DATA --hparams='{"rnn_layer_sizes":[50]}' --num_training_steps=20000 --eval

Run TensorBoard to view training results

tensorboard --logdir=/tmp/basic_rnn/logdir

Go to http://localhost:6006 to view TensorBoard dashboard.

Run training with script

Alternatively, there is a shell script included for your convenience. Run it from magenta/models/basic_rnn.

./run_basic_rnn_train.sh $EXPERIMENT_DIR $HYPERPARAMETER_STRING $NUM_TRAINING_STEPS $TRAIN_DATA [$EVAL_DATA]

Where

  • $EXPERIMENT_DIR is the experiment directory, such as /tmp/basic_rnn/logdir
  • $HYPERPARAMETER_STRING is a Python dictionary literal containing hyperparameters, such as '{"rnn_layer_sizes":[50]}'
  • $NUM_TRAINING_STEPS is an integer giving number of training iterations to run, such as 20000
  • $TRAIN_DATA is the path to the training dataset (a tfrecord file), such as /tmp/basic_rnn/sequence_examples/training_melodies.tfrecord
  • $EVAL_DATA, an optional argument, is the path to the eval dataset (a tfrecord file), such as /tmp/basic_rnn/sequence_examples/eval_melodies.tfrecord

This script automatically assigns new folders for new runs in each experiment directory. The eval job will optionally run if the path to the eval dataset is given. This script also runs TensorBoard.

Generating melodies

To generate, we need to load a checkpoint of a trained model. Look at your experiment directory, in this example /tmp/basic_rnn/logdir. Choose the best run under that experiment, in this example /tmp/basic_rnn/logdir/run1. basic_rnn_generate will look for the most recent checkpoint in that run.

The generator takes a melody as input to prime the model, meaning the model will be fed the primer melody first and then extend it. This primer should be a short monophonic melody stored in a MIDI file. You can make a MIDI file with any MIDI sequencer or digital audio workstation. If you do not have one, there are online sequencers such as https://onlinesequencer.net/. We provided an example primer called primer.mid.

You can choose how many outputs will be saved. Here we save 10.

Let's generate melodies that are 128 steps long.

# Provide a MIDI file to use as a primer for the generation.
# The MIDI should just contain a short monophonic melody.
# primer.mid is provided as an example.
PRIMER_PATH=<absolute path of your primer MIDI file>

bazel run //magenta/models/basic_rnn:basic_rnn_generate -- \
--run_dir=/tmp/basic_rnn/logdir/run1 \
--hparams='{"rnn_layer_sizes":[50]}' \
--output_dir=/tmp/basic_rnn/generated \
--num_outputs=10 \
--num_steps=128 \
--primer_midi=$PRIMER_PATH