-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not able to replicate results #4
Comments
Hello. We used around 60k anime images from probably 10-15 movies. I am afraid we can not shate the dataset since it is protected under copyright. The problem might lie in the learning rate, which was raised some during training. Try using a lower one than specified. I could also try to re-train the network in a couple of weeks or so. To confirm this issue. |
Hey, so can you atleast share what movies you used for making dataset and way frames were extracted from them. Then i can make the same dataset on my own. Please also specify the quality (360p or 720p and so on). P.S. Now I made some progress with the training after doing some changes in hyper parameter and so on. Now when I train, its exactly the other way around. The discriminator loss becomes very small after some epochs but the generator is not able to learn (errG just fluctuating around 2) the generator keeps on producing almost similar images as input. I have arrived at a conclusion that maybe batchsize is a problem: I have kept batchsize as 4 since anything above is not supported by my gpu (@256 img size). Do you support this argument or do you think there can be other issue? Finally i would request you to share the exact hyperparameters you used for a successful training and whether small batchsize is an issue and also the way you made your dataset(The movie names and how frames were extracted from them). It would be a huge help for my research. Thanks |
Hey can you upload the "trained_netD.pth" file. I am training on my own dataset, but I cannot continue without that file. It would be really helpful if you could share me that file. Thanks |
@rathodhare The hyperparameters should not be changed from the ones specified. I am missing the dataset myself atm. But I am acquiring it and can try training soon to confirm it still works. Edit: The movies used are:
|
I found a few bugs in the code but so far nothing seem to prevent training to achieve the results presented in the paper. @rathodhare did you pre-train the Generator at all? We did not have any file for this but is now available in this branch: https://github.com/FilipAndersson245/cartoon-gan/tree/replicationfixes The code is WIP and for now it trains on a static 10 epochs on the real dataset. I am currently training to make sure everything works as expected. |
Training for ~34 epochs seems successful. Epoch 0: Epoch 34: Trained using @rathodhare See if you can confirm or I'll close this issue in a few weeks. |
Hi,
Thanks for your reply. I wanted to mention that there is no file for the
pre-trained Discriminator in the repository. Thanks for providing the
pre-trained Generator file though. It would be really helpful if you could
provide the pre-trained Discriminator file as well.
Thanks
…On Mon, Mar 29, 2021, 11:25 PM zimonitrome ***@***.***> wrote:
Training for ~34 epochs seems successful.
Epoch 0:
[image: 0_0]
<https://user-images.githubusercontent.com/17656709/112878532-5b9ff280-90c8-11eb-9ecf-efb663d534da.png>
Epoch 34:
[image: 24_368]
<https://user-images.githubusercontent.com/17656709/112878572-66f31e00-90c8-11eb-9462-1ac46a6bcdc4.png>
Trained using pretrain.py and train.py.
@rathodhare <https://github.com/rathodhare> See if you can confirm or
I'll close this issue in a few weeks.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
<#4 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AIKFSXMCIRX4I6ON622RZPDTGDEIRANCNFSM4ZONN6YA>
.
|
So just to confirm, the current version of code would replicate the results with no changes in code in anyway by pretraining on Flickr30k (entire dataset?) and then training on 60k anime images from the movies you described and again same Flickr30k? Also how powerful is your GPU man, mine shows CUDA out of memory in batch size 4 cant hope to train on batchsize 32 .... Please confirm the above and upload the final trained Gen and Disc also the pretrained Gen if possible. Thanks for all your hard efforts :) |
Yes this is correct. You also need to process your anime images with We currently have access to an RTX 3090, hence the enormous batch size. You could experiment with higher learning rate for quicker results. The default is |
Hello, so i want to use your work for further research hence I am using your repository to replicate your results on my dataset. But while training, I am not able to get cartoonish image but the output looks nearly same to real image. The problem is that Discriminator is not training properly. It always give near 0 outputs (Real like patch prediction) for cartoon images (around 0.3~0.5 mean), and also for real images (~0.01) hence the generator is not able to train properly to yeild cartoon images.
I have followed procedure as you told and also as mentioned in the original Cartoon GAN paper: I pretrained generator to reproduce real images (6k images from Flickr30k dataset), for 10 epochs then I pretrained Discriminator as a normal classifier (6k images from flickr, 4.6k anime images from 3 movies of Hayao : PoppingHill, Princess Mononoke, Spirited away and 4.6k corresponding smooth images) for 50 epochs (still having same problem as i described earlier). After this I trained the combination for 50 epochs on the same data (6k images from flickr, 4.6k anime images from 3 movies of Hayao : PoppingHill, Princess Mononoke, Spirited away and 4.6k corresponding smooth images).
Can you suggest what is the problem? I have not made any changes to your code except for writing new codes for pretraining Generator and Discriminator. Maybe the problem lies in the dataset? or am I doing something else wrong? please help me out.
It would be great if you could share your dataset with us. Thanks.
The text was updated successfully, but these errors were encountered: