-
Notifications
You must be signed in to change notification settings - Fork 696
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Floating point exception (core dumped) #49
Comments
@ZohaibAhmed Unfortunately I don't see the same error on my end - can you do me a small favor? If you have an IDE with breakpoints can you check which function is causing that in gen_tacotron.py (should be somewhere in the loop starting on line 91)? If you don't have breakpoints you can just print('a', True), print('b', True) after each function in that loop to see what's throwing the error. Thanks. |
looks like the issue is on the vocoder generate function in fatchord_wavernn, specifically when it calls:
Note, that just using the pretrained model out of the box seems to work. It's just when I train the model further, the error occurs. More details about my setup: ubuntu16.04 |
@ZohaibAhmed can I get the exact steps you went through to get that error? Have you tried training a fresh model for a couple of epochs and then tried generating? Also is there no other error message besides "Floating point exception (core dumped)"? |
@fatchord - training a model from scratch seems to work. The exact steps I did were as follows:
And that's how I get to that error. Even if i keep the WaveRNN as the pretrained model, it still results in the |
@ZohaibAhmed can you try training LJ from scratch to see if you get the same error? |
@fatchord training Tacotron from scratch makes it work. But I don't have enough data for my own dataset to effectively train the model. Have you had any success with fine-tuning? EDIT: the main issue seems to be that the decoder is producing all silent values It looks like the shape of the output from the original pretrained model is different then when I train on top of it: Original: Tuned: Looks like I hit the condition where if silent frames are present:
This is what the alignment plot looks like while training tacotron: |
@ZohaibAhmed I met the same error. The reason is that the first frame of if (mel_frames < -3.8).all() and i > 10 : break |
@candlewill Nice catch, I'll push a fix for that later today. |
@candlewill - I still largely get silence (with some static). Did you try to train your model on top of the checkpoint that @fatchord provided? Or did you just train it from scratch? |
Tacotron has been updated to fix the premature stopping of generation. |
I tried to train the tacotron model you have on top of the LJ pretrained checkpoint you have. Just ran
train_tacotron.py
but when I rungen_tacotron.py
, I get the following:Any ideas on how I can go on debugging this?
The text was updated successfully, but these errors were encountered: