New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MASK LOSS to 0 #32
Comments
Hello,
|
yes, i use 256x256 and just fix some difference for the difference of the formats at Celeba.py. The output shows that mask ->0 and G_cls around 30 . |
I've tried to run for several epochs but this problem seems to never improve. |
May I know what kind of labels are you using for that dataset? |
Yes, i use 'Black_Hair', 'Blond_Hair', 'Brown_Hair', 'Male', 'Young', 'Smiling', 'Eyeglasses', 'Goatee', 'Pale_Skin', 'Bangs', the same setting of my model, i'm ready to cite your paper and do some comparison, so it's a little urge. |
Yes but as far as I am aware CelebA-HQ (at least the original implementation from here) does not have any labels. |
Can you send me a screenshot or something regarding the problem, so I can give you a more detailled feedback? |
Some one has collect the mapping from celeba-hq to celeba. So the labels are available somewhere. |
Actually, because of the different format, i used to run successfully once but the order of labels is wrong (csv begins at 1 but txt begins at 0, the first line of all labels). I fixed it and then the training begins to fail. |
Now I see. So it is not that the attention loss goes to zero but it remains saturated: if Normally it is a problem of random initialization that should be easily fixed by Please try this and let me know. |
Thank you, i'll run for one night to see if this problem can be solved. |
Hello, i've tried to run for one night, around 18 epochs with another seed, but this problem is still not improved. This time the G_cls is about 40 and D_cls about 4, why? |
I refreshed all files and try to find the reason. |
And i do think the attention loss is not what i thought it is. In Ganimation, the output is: |
I do not know what you are doing.
This is to ensure you are inserting a one hot encoding vector. If you are not inserting it, can I know what kind of labels are you using? If that was the problem it would have raised an error.
Why do you say so? This is exactly what we do as we mention it in the paper in section 3.2.1 and in the code here and here. I do not know if you were able to fix it. Without screenshots or more detailled instructions to reproduce it I cannot help you much. |
Yes, i know your code is doing what you show in paper, but this is different from the attention loss from other models, e.g. GANimation. In their papers, the change part should be minimized. |
Perfect. Let me know if anything good or bad happens and feel free to close the issue or keep it open until we fix this. |
Okay, i would feedback timely. And thank you so much for your timely reply, too. |
Hello, this time the results are good enough for me to compare. |
Thanks for your nice feedback. I will have that in mind from now on. |
I apply your code to celeba-hq, but found the loss of attention turn to 0 and so the networks do nothing. Why?
the d_cls is about 10 while g_cls go to 30 but networks seem to never optim this.
Another question is that why you set the parameter of DE fixed?
The text was updated successfully, but these errors were encountered: