Skip to content
forked from ciaua/unagan

Code for Unconditional Audio Generation with GAN and Cycle Regularization

License

Notifications You must be signed in to change notification settings

buganart/unagan

 
 

Repository files navigation

Unconditional Audio Generation with GAN and Cycle Regularization

This repository contains the code and samples for our paper "Unconditional Audio Generation with GAN and Cycle Regularization". The goal is to unconditionally generate singing voices, speech, and instrument sounds with GAN.

The model is implemented with PyTorch.

Paper

Unconditional Audio Generation with GAN and Cycle Regularization

Install dependencies

pip install -r requirements.txt

Download pretrained parameters

The pretrained parameters can be downloaded here: Pretrained parameters

Unzip it so that the models folder is in the current folder.

Or use the following script

bash download_and_unzip_models.sh

Usage

Display the options

python generate.py -h

Generate singing voices

The following commands are equivalent.

python generate.py
python generate.py -data_type singing -arch_type hc --duration 10 --num_samples 5
python generate.py -d singing -a hc --duration 10 -ns 5

Generate speech

python generate.py -d speech

Generate piano sounds

python generate.py -d piano

Generate violin sounds

python generate.py -d violin

Vocoder

We use MelGAN as the vocoder. The trained vocoders are included in the models.zip

For singing, piano, and violin, we have modify the MelGAN to include GRU in the vocoder architecture. We have found that this modification yields improved audio quality. For speech, we directly use the trained LJ vocoder from MelGAN.

Train your own model

One may use the following steps to train their own models.

  1. (Singing only) Separate singing voices from the audios you collect. We use a separation model we developed. You can use open-sourced ones such as Open-Unmix or Spleeter.

    wandb login # Only if not logged in yet.

    python scripts/collect_audio_clips.py --audio-dir "audio/RAW Sessions" --extension WAV python scripts/extract_mel.py python scripts/make_dataset.py python scripts/compute_mean_std.mel.py python scripts/train.hierarchical_with_cycle.py

Resume training

You need the wandb run id, then

python scripts/train.hierarchical_with_cycle.py --model-id 1r4b1ojr

Audio samples

Some generated audio samples can be found in:

samples/

Copy over trained melgan files

Note: these steps are from memory there might be steps missing or mistakes.

Assuming the model name custom the directory should look like this.

models/custom/
├── mean.mel.npy
├── params.generator.hierarchical_with_cycle.pt
├── std.mel.npy
└── vocoder
    ├── args.yml
    ├── modules.py
    ├── params.pt

Assuming melgan was trained with

python scripts/train.py --save_path checkpoints/harmonics --data_path data/harmonics

Assuming the melgan directory contains the melgan repo, the files to copy are

mkdir -p models/custom/vocoder
cp melgan/checkpoints/harmonics/args.yml models/custom/
cp melgan/mel2wav/modules.py models/custom/vocoder/
cp checkpoints/harmonics/best_netG.pt models/custom/vocoder/params.pt

The unagan model also needs to be moved, for example:

cp training_data/exp_data/{mean,std}.mel.npy models/custom
cp checkpoints/save/20200827_211112/model/params.Generator.latest.torch models/custom/params.generator.hierarchical_with_cycle.pt

Generation can then be done via

python generate.py --gid 1 --data_type custom --arch_type hc --duration 10 --num_samples 10

Todo

  • Train melgan-neurips.
  • Document how to combine the models for generation.

About

Code for Unconditional Audio Generation with GAN and Cycle Regularization

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 80.4%
  • Jupyter Notebook 18.9%
  • Other 0.7%