-
Notifications
You must be signed in to change notification settings - Fork 45.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Resource exhausted: OOM when allocating tensor with shape[2304,384] Traceback (most recent call last): #1993
Comments
The most expedient way is probably to reduce the batch size. It'll run slower, but use less memory. |
Are you saying that there is a memory leak? |
thanks for the answer |
thanks I reduced the batch size and it worked! |
Yay!! |
Im having the same issue |
I am using CPU to train the model. Please help |
I am having the same issue even with a reduced batch size |
same here. The batch size was already 1 and i've change the fixed_shape_resizer as 500x500 (using faster rcnn models) and
is also set. But it keep showing (by ssd resnet, same error) :
i'm using gtx 1060 6G and RTX 2070. both has same error |
Helllo, facing the same problem when training with the kangaroo data set . from Thanks for your help and advise .... |
the issue is there, not sure why the ticket is closed. There must be a leak since if you reboot the machine it goes away |
This issue should not be closed.. Tsk |
Hi, Which file to modify? |
@o92 I would say this is not a tensorflow issue to begin with. There are multiple things you can do:
|
these oom allocation errors main reason is the unavailability of enough ram so these are the fixes you can try
Reducing the batch size didn't help me first then I did this now it's working |
I encountered this issue when trying to do fine tuning on Colab with
|
Hi all,
As you can see in the error, there is OOM when allocating tensor with shape [8,1024, 4849,44]. My questions are:
Thanks a lot in advance. |
Coming here as reducing the batch size problem didn't solve in my case. If you are having this problem during inference, the following might help:
The difference between CPU and GPU inference time is not that high, and we'll have way more memory available using CPU. |
God bless you dude.. was struggling with this for 2 days |
Please go to Stack Overflow for help and support:
I tried to run models/tutorials/image/cifar10/train.py
I let it run about a day on my pc :
(windows10 , tensorflow-gpu 1.2 ,) after
2017-07-20 13:58:20.441224: step 941580, loss = 0.14 (3076.2 examples/sec; 0.042 sec/batch)
`I got this error :
`how can I fix it? and do I have to run it again from or the previous result is saved?
ibe the problem clearly here. Be sure to convey here why it's a bug in TensorFlow or a feature request.
The text was updated successfully, but these errors were encountered: