New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU util is not 100% #26
Comments
|
Also, I have a plan to implement all our implementations using DDP and try to fill out the missing part of our table after the camera-ready deadline. sincerely, Minguk Kang |
Hi @mingukkang, thank you for your quick response! It would be awesome if you provide the official DDP implementation. Thank you for your service! 🤩 |
Hi @sangwoomo:) Recently, I make the ddp branch, and it fully supports distributed data parallel (DDP) training. if you want to use it, plz follow the below constructions:
Thank you! I also plan to add syncBN using PyTorch's official synchronized BN module. I will merge the branch as soon as possible. Sincerely, |
Thanks a lot! :) |
I have updated all of StudioGAN's models to support DDP training. We can use PyTorch's official SyncBN and Mixed Precision Training to stabilize and speed up training GANs. Sincerely, Minguk Kang |
successfully done:) |
Hi, thank you for a great repository!
I'm trying to train SNGAN on my custom dataset (128x128) in 4 GPU, but found that the GPU utils are ~90%. I also increased
num_workers=16
and applied-rm_API
but it is still not addressed. Do you have any suspicion on this issue?By the way, even BigGAN_256 raised out-of-memory (OOM) error for 4 GPUs of 12G memory. Do I need 8 GPU server or distributed training for BigGAN_256?
Thank you for your help!
p.s. typo:
load_frameowrk
->load_framework
The text was updated successfully, but these errors were encountered: