Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

did anybody successfully train? #5

Closed
badend opened this issue Apr 1, 2016 · 1 comment
Closed

did anybody successfully train? #5

badend opened this issue Apr 1, 2016 · 1 comment

Comments

@badend
Copy link

badend commented Apr 1, 2016

with GPU I just reduced a lot of options such as batch size, feature_maps etc.

Always OOM occured!!

Resource exhausted: OOM when allocating tensor with shape[50,450537]
W tensorflow/core/common_runtime/executor.cc:1102] 0x8bcfce0 Compute status: Resource exhausted: OOM when allocating tensor with shape[50,450537]
[[Node: LSTMTDNN/LSTM/Linear_34/MatMul = MatMul[T=DT_FLOAT, transpose_a=false, transpose_b=false, _device="/job:localhost/replica:0/task:0/gpu:0"](LSTMTDNN/LSTM/dropout_34/mul_1, LSTMTDNN/LSTM/Linear/Matrix/read)]]
I tensorflow/core/common_runtime/gpu/pool_allocator.cc:244] PoolAllocator: After 2987 get requests, put_count=1597 evicted_count=1000 eviction_rate=0.626174 and unsatisfied allocation rate=0.833612

@carpedm20
Copy link
Owner

@badend Resource exhausted means your GPU doesn't have enough memory to build a graph which is needed for this repo. You'd better reduce the size of embedding or # of kernels.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants