Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Attention Blank, is it because my progressive training schedule? #223

Open
lilmadman007 opened this issue Jan 25, 2021 · 4 comments
Open

Comments

@lilmadman007
Copy link

lilmadman007 commented Jan 25, 2021

my attention is empty after 10k steps, which shouldn't be normal.
I'm using LJSpeech dataset.
This is the second time I preprocessed everything and trained.

image

Loss is around 1.0 at 10k steps
Are my settings wrong here? Does this not work?

image

Thanks!

NOTE: I LOOKED AT THIS ISSUE ALREADY -> #154

@fatchord
Copy link
Owner

fatchord commented Feb 1, 2021

Hi, sometimes the alignment will fail randomly. I've never tried with batch size of 8 so that could be it. Maybe try finetuning on one of the pretrained models.

@AhmadAlAmin21
Copy link

did you ever solve this?

@lilmadman007
Copy link
Author

did you ever solve this?

Sorry for the lack of feedback. No I did not. when fatchord commented that it fails sometimes I tried it again 2 more times, but it
just didn't work. Maybe my gpu is just not good enough, like I said, but I just moved on when I couldn't get results.
Any help would be appreciated anyways!

@AhmadAlAmin21
Copy link

I think i found a solution,

  1. increase "r" from 7 to 12 in the tts_schedule in hparams.py.
  2. go to models file>tacotron.py, and change line 200 from "scores = torch.sigmoid(u) / torch.sigmoid(u).sum(dim=1, keepdim=True)" to "scores = F.softmax(u, dim=1)".

got this from #154 (comment)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants