We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hi, I try to train your model on my personal datasets.
Are there any parameters you could suggest to tweak during training on my dataset? Or default params are enough for good quality results?
Are there any specific requirements to a dataset? I have the dataset of similar size like LJSpeech, 22050Hz, mono, similar duration distribution.
The text was updated successfully, but these errors were encountered:
Hey!
Thank you for your interest in 🍵 Matcha-TTS.
It should work just fine with the default parameters, but I've noticed that you can improve quality by adding more parameters to the decoder.
Matcha-TTS/configs/model/decoder/default.yaml
Lines 1 to 3 in 256adc5
Something like channel 512, 512 and 128 attention dim.
This seems to be good enough.
Sorry, something went wrong.
Thanks for your reply!
No branches or pull requests
Hi, I try to train your model on my personal datasets.
Are there any parameters you could suggest to tweak during training on my dataset? Or default params are enough for good quality results?
Are there any specific requirements to a dataset? I have the dataset of similar size like LJSpeech, 22050Hz, mono, similar duration distribution.
The text was updated successfully, but these errors were encountered: