So-to-Speak: an exploratory platform for investigating the interplay between style and prosody in TTS
Interspeech 2023 Demonstration
We introduce So-to-Speak, a customisable interface tailored for showcasing the capabilities of different controllable TTS systems. The interface allows for the generation, synthesis, and playback of hundreds of samples simultaneously, displayed on an interactive grid, with variation both low level prosodic features and high level style controls. To offer insights into speech quality, automatic estimates of MOS scores are presented for each sample. So-to-Speak facilitates the audiovisual exploration of the interaction between various speech features, which can be useful in a range of applications in speech technology.
The code implementation is using an adaptation of Nvidia's implementation of Tacotron 2 as a synthesis engine, HiFi-GAN as vocoder, and code and the trained model for the automatic MOS predictor.
Pre-trained models for the synthesis model and the HiFi-GAN model are provided as assets in the release and should be placed in the models/tronduo and models/hifigan folders respectively.
Run the script run_inference.py
from the repository should download both a base wav2vec model and a fine-tuned version, both of which are required to run MOS prediction script here. Both should be placed in the models/mos_ssl folder.
Run the code in the Jupyter Notebook SoToSpeak_launch_interface.ipynb to start the demo; availability of GPU and use of CUDA is required for synthesis.