Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

several questions about implementation details #2

Closed
cientgu opened this issue Jul 2, 2020 · 2 comments
Closed

several questions about implementation details #2

cientgu opened this issue Jul 2, 2020 · 2 comments

Comments

@cientgu
Copy link

cientgu commented Jul 2, 2020

Nice work! I have several questions about your paper:

  1. what' the detail setting about GAN and cGAN in Table3 and Figure 4. For cGAN, is the number of classess is 1000? what's the backbone of these two methods. Are they use all the imagenet images? is the pretrained model you released name "baseline" and "cgan" denotes "gan" and "cgan" in this table?
  2. How many images you used to calculate FID?(5k or 50k) Why the cGAN results much worse than Biggan, FID 35.14 compared with 7.4 reported in bigGAN. It's not even comparable but the visualization results seems good in your paper. How to explain it, is it because your diversity is much worse than BigGAN? Or some other explaination?
  3. How do you get the Logo-GAN results in Table 3? Did you re-implementate it? I could not find the results in their paper. Why do you think your results is a little bit worse than theirs?
  4. What you mean about "random labels" in table 3.

Thank you so much! I really appreciate your work.

@stevliu
Copy link
Owner

stevliu commented Jul 2, 2020

Hello,

Thanks for your interest in our work.

  1. All GANs we train on ImageNet use all the images of all 1000 classes. The backbone architecture is given by a ResNet backbone. And indeed, "baseline", where we train a vanilla GAN, corresponds to "gan".

  2. We use 50k images to calculate FID in all experiments. There could be several reasons our FID numbers are lower, since BigGAN uses a much larger architecture, and different discriminator regularization. This might result in lower diversity, which as you suggest could lead to the FID drop.

  3. We reimplemented Logo-GAN, and the details can be found in our appendix. Logo-GAN RC has indirect access to real labels by clustering a pretrained classifier, which could lead to better performance. On the other hand, Logo-GAN AE, which we outperform, is fully unsupervised.

  4. Random labels is when we fix a random labelling of the images, and train a GAN on the fixed labelling.

@cientgu
Copy link
Author

cientgu commented Jul 3, 2020

Thank you for your reply!

@cientgu cientgu closed this as completed Jul 3, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants