-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding autocomplete to vllm model.py #20
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM other than ask to make stream input optional
70b4547
11d5c96
to
70b4547
Compare
@tanmayv25 should we merge this? |
I was unable to get a successful run using this change with olga's pipeline. I am in middle of investigating what is not working. |
Seems like there was some setup problem. Tested successfully with job id 73883232. |
Does autocomplete work? It doesn't seem to be called anywhere in the code besides the method being defined? |
The auto-complete feature is currently available only on triton's main branch that tracks development to a future release. The auto-complete functionality will be available with 23.11 release which is scheduled to be released by end of the month. |
Moving default parameters from config to auto_complete function.
Also tests Python_backend's
set_model_transaction_policy