Skip to content

Hierarchical music transformer for symbolic generation.

License

Notifications You must be signed in to change notification settings

Egoluback/hierarchical_neuromusic

 
 

Repository files navigation

Hierarchical Music Transformer

Project description

Presentation about this project in general(in russian)

Presentation about Hierarchical Transformer(in russian)

This project provides implementation of Hierarchical Transformer(Hourglass) architecture for symbolic music generation. Music Transformer model with RPR self-attention is used as a basic transformer; it's implementation/training code based on Dmitrii Uspenskii project.

There were impemented various types of downsample/upsample functions from the paper and losses for different experiments. In all experiments model was training on Los Angeles MIDI Dataset 3.1 with REMI Tokenizer.

You can explore all the experiments on the WandB report.

Downsample/upsample functions:

  • Naive(reduce with mean/copy)
  • Linear(linear layer)
  • Attention resampling(attn function with naive/linear function)

Losses:

  • CrossEntropyLoss
  • CosineCrossEntropyLoss: the idea is to make pairs after shortening different $$- \sum_{i=1}^{n} y_i\cdot \log \hat{y_i} + \alpha\cdot \frac{1}{N} \sum_{k=1}^{N} \frac{1}{l_k} \sum_{i=1}^{l_k} \langle x_{l_k 2i}, x_{l_k 2i + 1}\rangle$$
  • ExpCosineCrossEntropyLoss: the idea is to encourage similarity to something already known for all pairs after shortening/upsampling $$- \sum_{i=1}^{n} y_i\cdot \log \hat{y_i} + \alpha\cdot \frac{1}{N} \sum_{k=1}^{N} \sum_{i=1}^{l_k} \sum_{j=1}^{l_k} e^{-\langle x_{l_k i}, x_{l_k j}\rangle}$$

Project structure

  • /scripts - project scripts
    • hierarchical transformer implementation is located in /scripts/model/Music_Transformer/hierarchical_music_transformer.py
    • new losses are located in /scripts/loss/CELossWrapper.py
  • install_dependencies.sh - script for dependencies installation
  • requirements.txt - Python requirements list
  • train.py - script to run train
  • test.py - script to run test

Hierarchical Transformer params

.json config for hierarchical transformer training is located in /scripts/configs/REMI/train_hierarchical_music_transformer.json

The most changeable params are:

  • input length: the amount of tokens on the first layer = n
  • shorten factor: shorten factor s from the paper
  • depth: (x, (y, z, y), x) means x layers with n tokens, y layers with n/s tokens, z layers with n/s^2 tokens, ...
  • attn resampling: whether to use attn resampling or not
  • updown sample type: type of down/up sample layer. now there are only "linear" and "naive"
  • save_updown_hidden: if set to true, upsample tokens will not be used for hidden loss calculations (this attribute is needed for CosineCELoss)

Reproduce results

To run train Music Transformer with REMI tokenizer and Los Angeles MIDI dataset:

python -m train -c scripts/configs/REMI/train_music_tranformer.json

To run test inference with Los Angeles MIDI dataset with 512 prompt tokens and generate 512 tokens:

python test.py \
   -c scripts/configs/test_LAMD.json \
   -r best_model/model_best.pth \
   -o test_results_LAMD \
   --prompt_length 512 \
   --continue_length 512 \
   -b 1

To test model on a custom dataset you need to put MIDI files in some directory. To run test with custom dataset in custom_dataset directory:

python test.py \
   -c scripts/configs/test_custom.json \
   -r best_model/model_best.pth \
   -o test_results_custom \
   --prompt_length 512 \
   --continue_length 512 \
   -b 1 \
   -t custom_dataset/

About

Hierarchical music transformer for symbolic generation.

Topics

Resources

License

Stars

Watchers

Forks

Languages

  • Python 99.8%
  • Shell 0.2%