Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Usage]: Vllm inference slower for LoRA models #3979

Closed
akrish2011 opened this issue Apr 10, 2024 · 2 comments
Closed

[Usage]: Vllm inference slower for LoRA models #3979

akrish2011 opened this issue Apr 10, 2024 · 2 comments
Labels
usage How to use vllm

Comments

@akrish2011
Copy link

Your current environment

The output of `python collect_env.py`

How would you like to use vllm

when running LoRA trained models using Vllm see lower inference speed when compared to Non-LoRA trained models. Is there anything causing this ?

@akrish2011 akrish2011 added the usage How to use vllm label Apr 10, 2024
@jeejeelee
Copy link
Contributor

jeejeelee commented Apr 11, 2024

please refer to: #2829 and #1804

@mgoin
Copy link
Sponsor Collaborator

mgoin commented Aug 2, 2024

This should be lessened by about 2x with the new landed Triton kernels #5036

@mgoin mgoin closed this as completed Aug 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
usage How to use vllm
Projects
None yet
Development

No branches or pull requests

3 participants