-
-
Notifications
You must be signed in to change notification settings - Fork 10.6k
Closed
Description
Model: Llama-2-chat-hf
The current implementation of vLLM gives the finish_reason as 'length' whilst the native model supports context length of 4024(And works well with the context we've tested it with) , Is the option available to change the native context length supported by the vLLM instance?
I've retried the experiments with the latest release and the issue still persists.
horiacristescu and SitwalaM
Metadata
Metadata
Assignees
Labels
No labels