RuntimeError with CUDA 12.2 on Windows using vLLM and Llava: No Kernel Image Available #4398
Closed
OualidBougzime
started this conversation in
General
Replies: 1 comment
-
I'm facing a similar issue for a different project and found this to be insightful. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I installed vLLM using the following command:
!pip install vllm==0.4.0 kaleido python-multipart torch==2.1.2 torchvision==0.16.2 torchaudio==2.1.2
Below is the code I am using with Llava:
However, I am encountering the following error:
RuntimeError: CUDA error: no kernel image is available for execution on the device
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with
TORCH_USE_CUDA_DSA
to enable device-side assertions.Versions:
CUDA: 12.2
OS: Windows
Python: 3.12.2
Any insights or suggestions to resolve this error would be greatly appreciated.
Beta Was this translation helpful? Give feedback.
All reactions