-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What is the configuration of your computer, such as GPU model and GPU memory size #430
Comments
2080ti, 11G |
If I want to train D7, how many GPUs(2080ti, 11G) should I prepare? |
I believe that you would not be able to fit the D7 model on 2080 Ti. It does not matter how many as D7 would not fit into RAM of a single 2080Ti (and you need to be able to load whole model into RAM of each GPU) even with batchsize of 1, unless you
I have a 2080 Ti and I stumbled at D5 |
@NikZak It seems training in FP16 is much harder for effdet based on my previous experiments. You can modify the train.py following the coco_eval.py to train in fp16. BTW, fp16 is not supported when using dataparallel. |
Just investigating the memory issue further:
However it can be trained on TPU without same memory issues
Obviously Google cares about TPU and nuances of the code could be different. Someone suggested that the trouble is caused by lines like this: For Tensorflow, there is a probable solution here (aggregation_method=tf.AggregationMethod.EXPERIMENTAL_ACCUMULATE_N) but it does not seem to be working for EfficientDet implementation as people are still complaining. However a solution must exist as training (of the Tensorflow official model) works on TPU |
Thank you for your amazing job!
What is the configuration of your computer, such as GPU model and GPU memory size?
I'm looking forward to your reply.
The text was updated successfully, but these errors were encountered: