-
-
Notifications
You must be signed in to change notification settings - Fork 6.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: can't pickle model config error on V0 engine for deepseek-r1 #14966
Comments
The same model runs okay with |
See #14925 |
@cjackal can you please share the whole CLI command which reproduces this error, including (but not limited to) the specific deepseek r1 model you used? Thanks |
For example I was able to successfully run this command VLLM_USE_V1=0 vllm serve --trust-remote-code deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B --port 8091 so I am wondering if this is not the model you used |
Based on the other issue, perhaps the error only occurs when directly referencing a local model via filepath? |
Deepseek distill qwen models belong to Qwen2 model architecture, you may try a deepseekV2 model architecture for reproducibility. I think DeepSeek-V2-Lite is the smallest. What I'm using is https://huggingface.co/unsloth/DeepSeek-R1-GGUF/tree/main/DeepSeek-R1-UD-Q2_K_XL, but as GGUF models must first be joined you need to first download the model binary to local filepath and follow this. The specific launch command should look like: VLLM_USE_V1=0 vllm serve /app/model/deepseek-r1-gguf-q2-k-xl/DeepSeek-R1-UD-Q2_K_XL.gguf --served-model-name deepseek-ai/deepseek-r1 --tokenizer /app/model/deepseek-r1-gguf-q2-k-xl/ --hf-config-path /app/model/deepseek-r1-gguf-q2-k-xl/ --trust-remote-code ... |
Your current environment
AFAICT it happens regardless of the GPU arch
the particular version that I have tested is:
vLLM API server version 0.8.0rc2.dev9+g6eaf1e5c (for cuda)
🐛 Describe the bug
When
VLLM_USE_V1=0
, launching vLLM server with--trust-remote-code
fails with the following error -Can't pickle <class 'transformers_modules.configuration_deepseek.DeepseekV3Config'>: it's not the same object as transformers_modules.configuration_deepseek.DeepseekV3Config
traceback
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: