-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Issues with CUDA_OUT_OF_MEMORY #132
Comments
Hi, |
Hi I have same issue here:
no matter which nmt size I choose: |
Hi, E tensorflow/core/common_runtime/direct_session.cc:137] Internal: failed initializing StreamExecutor for CUDA device ordinal 0: Internal: failed call to cuDevicePrimaryCtxRetain: CUDA_ERROR_OUT_OF_MEMORY; total memory reported: 18446744073709551615 Have tried the earlier suggested fix of resetting the flag " gpu_allow_growth" to True. Kindly suggest. Thanks |
Have you tried CUDA_VISIBLE_DEVICES flag while calling train/test.py? The problem is, you are not able to create session. |
Hi
When trying out the pipeline unit test I found here, I got the two following errors:
As there is nothing running on the two GPU's (I checked with nvidia-smi) and I am the only person at my internship to try stuff out on them, I don't find a reasonable explanation. Could someone point me in the right direction? As I'm a newbie to tensorflow and GPU's in general, I find it hard to know where to start.
Thanks in advance
The text was updated successfully, but these errors were encountered: