New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
probav training memory error #24
Comments
I trained the model in the example on Colab with a Tesla T4 so maybe it gave you a different GPU? Anyways, you cam decrease the batch_size and then increase accumulate_grad_batches to deal with this. Alternatively, you can reduce the num_res_blocks in the model. |
Strange. There must be a memory leak somewhere mayne in pytorch lightning? I'll look into this. |
I had Tesla P100. Will try your suggestions.
|
|
Increasing from 5 to |
Using colab pro with nominally 25 Gb I am still running out of memory at 17 epochs using your probav example notebook. Is there any way to free memory on the fly? I was able to train the tensorflow RAMS implementation to 50 epochs on colab pro
The text was updated successfully, but these errors were encountered: