-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About batch_size in training configuration? #46
Comments
Hi @xiao2mo, there is no special purpose. We set batch size 32 per GPU because some of our machines are 16G per card, and we need to test some other parameters like frame number. It is an appropriate batch size to finish the hyper-parameters study. If your card has more than 16G, a suggestion is to test a large frame number, then a large batch size. |
I see it. The main problem in my exps is that batch size in DDP configuration may result in different results. Thank u. |
hi 训练配置如下
运行命令如下
|
Hi Luo, thanks for your remarkable work!
I am here wondering why use batch 128 in 4gpu DDP training? That is to say 32 per GPU which only accounts for less than 1/2 GPU memory. Is that for some special purpose?
The text was updated successfully, but these errors were encountered: