-
Notifications
You must be signed in to change notification settings - Fork 4.7k
vLLM inference servers container image fixes #582
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
/assign @janetkuo |
| # 1 billion parameter model (smallest gemma model) | ||
| - name: MODEL_ID | ||
| value: google/gemma-3-1b-it | ||
| # Fixes - https://github.com/vllm-project/vllm/issues/18859 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is the information here meaningful to users? Is LD_LIBRARY_PATH only needed for certain version of vLLM? If so, that kind of information is good to capture here.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good point. I've added the versions of the image that this is needed for. I've kept a reference to the issue, in case users need to do more research. Please let me know what you think.
7cbcf02 to
a0ed769
Compare
janetkuo
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
/lgtm
|
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: janetkuo, seans3 The full list of commands accepted by this bot can be found here. The pull request process is described here
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
v0.11.0).LD_LIBRARY_PATHenvironment variables to make the server more robust. This environment variable can be missing from some container images.