Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Get error when compiling. #25

Closed
Cortega13 opened this issue Jun 3, 2023 · 2 comments
Closed

Get error when compiling. #25

Cortega13 opened this issue Jun 3, 2023 · 2 comments

Comments

@Cortega13
Copy link

Cortega13 commented Jun 3, 2023

Hello! I am trying to run exllama on wizard-vicuna-13b-uncensored-gptq, and when i try to run any of the commands I get the following error. I am running it using the nvidia pytorch image nvcr.io/nvidia/pytorch:23.05-py3. I am using the newest version of cuda 12.1.1 and running it on a google vm with an L4 on ubuntu 18.04 LTS. I know the documentation says its not compatible with all gpu's, is it compatible with the L4? Any help would be very much appreciated. Thank you!!

error.txt

@turboderp
Copy link
Owner

I don't know anything about the L4, sadly. But the error you're getting seems to be because for some reason it's targeting some very old compute versions. Perhaps the PyTorch image is just set up like that. But you could try this before running it:

export TORCH_CUDA_ARCH_LIST="sm_89"

The L4 should support compute 8.9, I think. Then at least you shouldn't get those errors. But whether it will actually work and what the performance will be, no idea. It's a low-bandwidth, "energy efficient" GPU by the looks of it.

@Cortega13
Copy link
Author

Thank you!!! The command was export TORCH_CUDA_ARCH_LIST="8.9"

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants