Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cuda out of memory #3

Closed
winsapdev opened this issue Apr 29, 2022 · 2 comments
Closed

Cuda out of memory #3

winsapdev opened this issue Apr 29, 2022 · 2 comments

Comments

@winsapdev
Copy link

hi Jaketa

thanks for making this code. anyway i run on colab and got error when parsing torch tensor to the model
out memory

can you help me with this error. thank you very much

Regards,
Winsap

@jaketae
Copy link
Owner

jaketae commented Apr 29, 2022

Hey @winsapdev, seems like you ran into a CUDA out of memory on Colab. CUDA OOM is normally triggered because the model you're dealing with is too large to be dealt with by the GPU. You can most likely resolve it by declaring a smaller model. In this case, try decreasing d_model or d_ffn. Alternatively, if you have access to a larger GPU, you can try out the model there. Hope this helps!

@jaketae
Copy link
Owner

jaketae commented Jun 9, 2022

Closing this for now, feel free to ping me if you have any questions!

@jaketae jaketae closed this as completed Jun 9, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants