-
Notifications
You must be signed in to change notification settings - Fork 573
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can pascal ever be supported? Is it a lost cause? #131
Comments
This is supported in the latest release. Please update to 0.37.0 and let me know if you are still having troubles. |
@TimDettmers Maxwell user here (Nvidia M40), when trying to import the library I get the following error:
|
Thank you for fixing it so quickly. Unfortunately I have the same problem as henk does above. In addition to trying kobold, I used https://github.com/oobabooga/text-generation-webui and got the same "NameError: name 'cuda_setup' is not defined" Also this additional error: https://pastebin.com/9wySYKfv When I tried searching this was related to it not being able to find the cuda libs? |
The same environment I use for Kobold works on a newer GPU, so I think the not finding the libs error is deceiving. |
I am able to bypass the error by doing this:
But now I get an error like this: https://pastebin.com/H0gvKSMF when I try to generate text. |
KoboldAI isn't going to be a good example because we do things quite different from other projects and have no official 8-bit support yet. I recommend using something else first based on more basic HF functions. |
Any suggestions? I also had the same failure in the textgen-webui that natively supports 8 bit. From looking it up, it's some kind of divide by zero error due to torch. I got a new error message and attempt at generation. Turned off stuff like temperature/ top P. Acts like it generated but get nothing except exclamation marks and this.
|
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
I have a quadro built on pascal and see a bunch of people with P40 posting here. Someone has compiled the library for windows and is using it on a 10xx series geforce for training.
I see some docs and forum post from nvidia.
https://developer.nvidia.com/blog/new-pascal-gpus-accelerate-inference-in-the-data-center/
https://forums.developer.nvidia.com/t/about-cublasgemm-int8-support/48483
But the problem is that int8 multiplication function.
Is it a hard blocking issue that we should give up hope on? The most accessible high vram GPUs are in this series.
Also, since they are indeed running this for training, does this mean other 8 bit operations are currently working and it's just the one required for language models?
The text was updated successfully, but these errors were encountered: