Skip to content
AutoVC: Zero-Shot Voice Style Transfer with Only Autoencoder Loss
Branch: master
Clone or download
Latest commit 2d8a6c8 May 16, 2019
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
LICENSE 1st commit May 14, 2019
README.md update readme May 16, 2019
conversion.ipynb 1st commit May 14, 2019
hparams.py 1st commit May 14, 2019
metadata.pkl 1st commit May 14, 2019
model_vc.py 1st commit May 14, 2019
synthesis.py c is already numpy May 16, 2019
vocoder.ipynb 1st commit May 14, 2019

README.md

AUTOVC: Zero-Shot Voice Style Transfer with Only Autoencoder Loss

This repository provides a PyTorch implementation of AUTOVC.

AUTOVC is a many-to-many non-parallel voice conversion framework.

To ensure respect for privacy rights and responsible use of our code, we are only releasing a portion of our code to allow users to convert voices among a predefined set of speakers in VCTK. Conversions from and to other voices have been disabled.

Audio Demo

The audio demo for AUTOVC can be found here

Dependencies

  • Python 3
  • Numpy
  • PyTorch >= v0.4.1
  • TensorFlow >= v1.3 (only for tensorboard)
  • librosa
  • tqdm
  • wavenet_vocoder pip install wavenet_vocoder for more information, please refer to https://github.com/r9y9/wavenet_vocoder

Pre-trained models

AUTOVC WaveNet Vocoder
link link

0.Converting Mel-Spectrograms

Download pre-trained AUTOVC model, and run the conversion.ipynb in the same directory.

1.Mel-Spectrograms to waveform

Download pre-trained WaveNet Vocoder model, and run the vocoder.ipynb in the same the directory.

You can’t perform that action at this time.