Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about training from scratch #31

Open
Levent9 opened this issue Mar 24, 2024 · 2 comments
Open

Questions about training from scratch #31

Levent9 opened this issue Mar 24, 2024 · 2 comments

Comments

@Levent9
Copy link

Levent9 commented Mar 24, 2024

Hello, I followed the steps in run.sh to train with the LibriTTS-R dataset. Below is the training process loss. When I use the current checkpoint to synthesize speech, it is almost noise. Based on the loss, does the training of the model appear normal? Thank you!

Loss:
loss
nll_loss:
nll_loss
reg_l1_loss:
reg_l1_loss
reg_l2_loss:
reg_l2_loss

@ZhihaoDU
Copy link
Collaborator

I think the training process is normal. On LibriTTS corpus, I got losses like these after about 106K iterations:

loss=6.557
nll_loss=4.271
reg_loss=2.286
reg_l1_loss=1.733
reg_l2_loss=2.840
out_acc_1=0.167
out_acc_2=0.100

If the synthesized speech is almost noise, there may be other bugs ?

@Levent9
Copy link
Author

Levent9 commented Apr 8, 2024

Thank you very much for your response. The text-to-speech synthesis is now functioning normally.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants