Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

What do the output letters signify? #26

Closed
iboates opened this issue Mar 8, 2020 · 7 comments
Closed

What do the output letters signify? #26

iboates opened this issue Mar 8, 2020 · 7 comments

Comments

@iboates
Copy link
Contributor

iboates commented Mar 8, 2020

What do the letters "G", "D", "GP" and "PL" in the output signify?

0% 0/90000 [00:00<?, ?it/s]G: 1.51 | D: 0.21 | GP: 0.00 | PL: 0.00
0% 49/90000 [02:46<81:43:57,  3.27s/it]G: 2.06 | D: 0.05 | GP: 0.50 | PL: 0.02
0% 97/90000 [05:04<71:55:48,  2.88s/it]G: 1.63 | D: 0.00 | GP: 0.00 | PL: 0.03
0% 149/90000 [07:35<72:52:45,  2.92s/it]G: 1.04 | D: 0.12 | GP: 0.00 | PL: 0.03
@lucidrains
Copy link
Owner

lucidrains commented Mar 8, 2020

@iboates you are back! any interesting training results? :)

Those are the vital signs of the training, the numbers that the network is forced to try to minimize.

G: generator loss
D: discriminator loss
GP: gradient penalty loss
PL: path length regularization loss

G and D are fighting each other, and ideally stay flat. When D hits 0 consistently, training is usually done, and the best generator is a few saved models behind. GP should be 0 for stability. PL will occasionally spike as G learns something new, but should ideally be pushed back close to 0.

@iboates
Copy link
Contributor Author

iboates commented Mar 8, 2020

Thanks for the info.

The results so far have been disappointing unfortunately. I have been trying to train it on images of maps coming from openstreetmap.

It was making good progress, but after about 70k iterations the model collapsed and began outputting random smears of colour. I was thinking that maybe it had to do with the fact that the data pool was pretty small, and what images I did have had quite a bit of variety.

So I have come up with a way to get much more data, and isolated the maps so that they are always featuring villages or small towns. I'm training again with about 3k of these images, but I can generate many many more. But right now I think I am hitting the upper end of Google Colab, it takes about 7 hours to do 10k iterations, and that is when the checkpoint is made. Since Colab disconnects after 10 hours, I can't really squeeze more data in. I think I have to buy some cloud processing time or something.

This was about as good as it got before collapsing:

15

@lucidrains
Copy link
Owner

@iboates ahh 3k training set is nothing! you need up to 10k or more!

@iboates
Copy link
Contributor Author

iboates commented Mar 8, 2020

Thanks for the clarification. We're a bit off-topic from the original question, but how many faces did you train it on for thispersondoesnotexist.com? And do all the images have to be the same size or can they just be of similar size & resolution to depict what they are showing?

@lucidrains
Copy link
Owner

@iboates to get to that level of quality, I would recommend the official repository, as it is more optimized. That model was trained on 70k high quality images by Nvidia

@iboates
Copy link
Contributor Author

iboates commented Mar 8, 2020

Do all the images need to be the exact same size?

@lucidrains
Copy link
Owner

@iboates ideally yes!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants