-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
1 RTX A6000 out of memory #5
Comments
One possible solution that works for me is to use 4 RTX A6000 (3$ an hour). The setup was successful there and could start fine tuning. Might also be possible to use "1 RTX A6000" without running out of memory by using other docker images among these
|
Hi, I haven't tested with other images rather simple default cuda image with current setup. Using Pytorch image might be the reason the disk is out of memory. Furthermore, it's not necessary to have pytorch image since the environment already has one. I think I have tested with cuda:12.0.1-runtime-ubuntu20.04 |
Trying as suggested in README.md to run this on "1 RTX A6000" with docker image
pytorch:latest f5540ef1a1398b8499546edb53dae704
from https://cloud.vast.ai/Returns out of memory error
Debugging memory
The text was updated successfully, but these errors were encountered: