Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix: patch vllm/local endpoint model GET bug #1179

Merged
merged 1 commit into from
Mar 21, 2024
Merged

Conversation

cpacker
Copy link
Owner

@cpacker cpacker commented Mar 21, 2024

Please describe the purpose of this pull request.

Model listing during CLI was broken for local endpoints because for our local model GETs (to generate the list): we assume the server supports the OpenAI-proxy style of GET _/v1/models, however we don't store local model endpoints with the /v1 suffix since many of the local LLM inference servers don't use it.

This fixes the issue by appending a /v1 in the model GET call if it doesn't exist (since on this call we're assuming OpenAI proxy anyways).

Working as intended:
image

@cpacker cpacker merged commit b199573 into main Mar 21, 2024
4 checks passed
@cpacker cpacker deleted the model-listing-patch branch March 22, 2024 19:36
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
Status: Done
Development

Successfully merging this pull request may close these issues.

None yet

1 participant