You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @bibutikoley, vLLM doesn't support models in the GGUF format. Please use original precision models or one of the many quantizations we do support, such as GPTQ, AWQ, FP8, etc. Thanks!
Your current environment
I tried running the vLLM using the TheBloke/Mistral-7B-Instruct-v0.1-GGUF with the below command
But got the below error.
OSError: It looks like the config file at '/aimodels/mistral-7b-instruct-v0.1.Q4_K_S.gguf' is not a valid JSON file.
How would you like to use vllm
I would like to use the TheBloke/Mistral-7B-Instruct-v0.1-GGUF using vLLM
The text was updated successfully, but these errors were encountered: