-
Notifications
You must be signed in to change notification settings - Fork 620
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Out of memory during training #304
Comments
I am using GeForce GTX 1080. |
you can run |
I'm having the same issue - has nothing to do with the batch size, GPU memory keeps increasing regardless. I am using |
@oguzelibol how have you checked this has nothing to do with batch size? |
I'm having the same issue too. I am also using CUDA8 and pytorch 0.4.0 with Python 3.5. Any ideas? Ihave set batch-size=1, but GPU memory keeps increasing. I use TITAN XP. |
@oguzelibol Maybe it is useful to reinstall pytorch and warp-ctc. |
@oguzelibol Setting the Docker |
@SeanNaren - Yes I have experimented with several different batch sizes and had the same issue. Here is the solution that worked for me (and worked regardless of the batch size) - hopefully this will also tell something about the root cause: -Roll back to pytorch 0.3.1 |
I'm having the same issue when training. It may be related to pytorch and RNN. Are there any solutions to it? @SeanNaren |
Same issue here on Pytorch 1.0.0 and latest warp-ctc with latest pytorch audio. Cuda goes OOM irrespective or layer dimensions or batch size. |
still same issue here on latest pull of repo with Pytorch 1.2 and latest warp-ctc. Cuda goes OOM irrespective or layer dimensions or batch size. |
@rajeevbaalwan have you tried with a batch size of 1? How much VRAM (GPU memory) do you have? |
@SeanNaren i am using 2080Ti having 11 GB memory |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
I am running out of memory on every epoch:
I have merged A4 and TED datasets and trying to train on the merged dataset and I am getting out of memory every epoch:
Is any way to set max process gpu memory in pytorch similar to TF:
Fortunately I am able to resume using checkpoints. Seems relevant to issue #172
The text was updated successfully, but these errors were encountered: