Vocos: Closing the gap between time-domain and Fourier-based neural vocoders for high-quality audio synthesis
Unofficial PyTorch implementation of Vocos by Hubert Siuzdak.
Vocos is a fast neural vocoder designed to synthesize audio waveforms from acoustic features. Trained using a Generative Adversarial Network (GAN) objective, Vocos can generate waveforms in a single forward pass. Unlike other typical GAN-based vocoders, Vocos does not model audio samples in the time domain. Instead, it generates spectral coefficients, facilitating rapid audio reconstruction through inverse Fourier transform.
Audio samples and some information are provided in the web.
Clone the repository and install dependencies.
# the codebase has been tested on Python 3.10 with PyTorch 2.0.1 binaries
git clone https://github.com/LqNoob/vocos
pip install -r requirements.txt
Prepare a filelist of audio files for the training and validation set. Fill a config file with your filelist paths and start training with:
python train.py -c configs/vocos-fs2.yaml
If you need to do fine-tune on the model. Run the following command:
python train.py -c configs/vocos-fs2-gta.yaml
Refer to Pytorch Lightning documentation for details about customizing the training pipeline.
Inference only supports the copy synthesis process, see:
python inference.py