-
-
Notifications
You must be signed in to change notification settings - Fork 8.8k
Closed as not planned
Labels
bugSomething isn't workingSomething isn't workingstaleOver 90 days of inactivityOver 90 days of inactivity
Description
Your current environment
There is memory leak when I set max_tokens=1
I have tested 2 versions:
pip install vllm==0.6.6
or
pip install vllm==0.7.2
If I call generate with SamplingParams(max_tokens=1), there is a memory leak.
🐛 Describe the bug
from vllm import LLM, SamplingParams
llm=LLM(model='qwen/Qwen2.5-0.5B-Instruct', trust_remote_code=True)
sampling_params = SamplingParams(max_tokens=1)
while True:
result = llm.generate("Hello world, ", sampling_params)
pass
pass
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't workingstaleOver 90 days of inactivityOver 90 days of inactivity