-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
vllm backend faild #2028
Comments
Hi, in order to help solve your issue we'd need more information. Namely:
|
full traceback , it's too long, abstract key lines: library version: vllm=0.5.0.post1, vllm-flash-attn =2.5.9, torch2.3.0+cu121) , lm-eval 0.4.2 |
I see the problem-- |
thank for ur comment, another error, test: and when pip install lm_eval[vllm], and fix refer #1953 |
Hi, could you try the following:
|
@chunniunai220ml are you by chance running your PyPI-installed |
at first , pip install lm_eval[vllm], error as reported. then git pull the latest code, pip install -e . , same error. |
hi, i tried to eval :
export CUDA_VISIBLE_DEVICES="2,3" accelerate launch -m lm_eval --model vllm \ --model_args pretrained="THUDM/glm-4-9b",dtype=bfloat16 \ --tasks mmlu \ --device cuda \ --batch_size 2 \ --trust_remote_code \ --cache_requests true \ --num_fewshot 5 \
for accelerate, but failed in A100 (vllm=0.5.0.post1, vllm-flash-attn =2.5.9, torch2.3.0+cu121):
The text was updated successfully, but these errors were encountered: