-
Notifications
You must be signed in to change notification settings - Fork 73
Troubles on running code with single GPU (RTX 2070 SUPER) and 16 GB RAM #12
Comments
Thanks for using the code. Are you using tf 1.15? I met the OOM issue at that version. Pls follow closely of the environment requirements. |
Hi,
I’m using tensorflow 1.14-gpu and following all of the instructions.
…On Tue, Aug 25, 2020 at 6:58 PM zzz ***@***.***> wrote:
Thanks for using the code.
Are you using tf 1.15? I met the OOM issue at that version. Pls follow
closely of the environment requirements.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#12 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ALUCCHUWZOAHWTZKQKALLCDSCQ62HANCNFSM4QK6L6NQ>
.
|
What GPU type do you use? It requires 16GB mem and we tested on V100 |
Hi,
I’m using rtx2070 super which only has 8GB RAM. I think that’s the problem.
…On Tue, Aug 25, 2020 at 7:21 PM zzz ***@***.***> wrote:
What GPU type do you use? It requires 16GB mem and we tested on V100
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#12 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ALUCCHVPO6OHNJ5L652KM4LSCRBQNANCNFSM4QK6L6NQ>
.
|
Hi, |
No it can not, we already use batch size = 1. If you are doing experiments on your own datasets and do not care about COCO benchmarks, you can consider reduce the input image size. |
@zizhaozhang |
@Chen-Song I do not think you can train and get reasonable results using 1 gpu. Since the batch size per GPU is only 1. Too small total batch_size will make network hard to converge. |
Thanks, how many GPUs do you use, and how long does it take to get an experimental result in COCO? |
@sisrfeng Each GPU needs to have 16 RAM. @Chen-Song We train on 8 V100 and it takes around 8-10 hours. |
Hello! I was trying to run the code but my process was always being killed by system which said 'out of memory'.
I already set batch size ==1, so what else can I do to run the code?
The text was updated successfully, but these errors were encountered: