-
Notifications
You must be signed in to change notification settings - Fork 18
Add support for api_key #116 #120
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: andresC98 The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
Welcome @andresC98! |
Please resolve the merge conflict |
Type of change
Description
This pull request introduces API key authentication for the
vLLMClient
. This allowsinference-perf
to connect to vLLM server instances that are protected and require anAuthorization
header for requests.Changes Implemented
inference_perf/client/modelserver/vllm_client.py
:__init__
method ofVLLMClient
has been updated to accept an optionalapi_key
argument.process_request
method, anAuthorization
header with a Bearer token is now added to the request if anapi_key
was provided during client initialization.inference_perf/main.py
:VLLMClient
is now instantiated with theapi_key
from the server configuration (config.server.api_key
), enabling the feature to be configured directly from the main config file.This change is backward-compatible and will not affect connections to vLLM servers that do not require authentication.