Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

openvino support in vllm #65

Merged
merged 11 commits into from
Jun 7, 2024

Conversation

zahidulhaque
Copy link
Contributor

Description

This PR adds support for openvino as a backend inference engine to the vllm.

Type of change

  • Add script to build Docker image and instructions for building it.
  • Provides examples for starting the vLLM serving container with OpenAI API endpoint and interacting with it via bash shell
  • Includes additional server start-up parameters and a curl example for requesting completion with vLLM

comps/llms/vllm-openvino/build_vllm_openvino.sh Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/launch_model_server.sh Outdated Show resolved Hide resolved
zahidulhaque and others added 3 commits May 17, 2024 11:56
* Adds  Docker image and instructions for building it
* Provides examples for starting the vLLM serving container with OpenAI API endpoint and interacting with it via bash shell
* Includes additional server start-up parameters and a curl example for requesting completion with vLLM

Signed-off-by: Zahidul Haque <zahidul.haque@intel.com>
Signed-off-by: Zahidul Haque <zahidul.haque@intel.com>
for more information, see https://pre-commit.ci

Signed-off-by: Zahidul Haque <zahidul.haque@intel.com>
@hshen14
Copy link
Collaborator

hshen14 commented May 17, 2024

@zahidulhaque please get all tests passed. Right now some typo issue was detected.

@lvliang-intel
Copy link
Collaborator

image
Please fix pre-commit issue.

Copy link

@mkbhanda mkbhanda left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I would add to the README motivation for using OpenVINO backend inference service, a few sentences like the runtime generates hardware target specific optimizations...

comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/launch_model_server.sh Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/launch_model_server.sh Outdated Show resolved Hide resolved
comps/llms/vllm-openvino/README.md Outdated Show resolved Hide resolved
zahidulhaque and others added 4 commits May 31, 2024 09:48
Copy link

@mkbhanda mkbhanda left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good!

Copy link

@mkbhanda mkbhanda left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Noted changes pertaining to License, hugging face token rename, service name change, and handling of results. Trusting your tests. Next time please strive for smaller PRs! Easier to digest all round.

@chensuyue
Copy link
Collaborator

Please contribute an e2e test for this microservice, like this one: https://github.com/opea-project/GenAIComps/blob/main/tests/test_reranks.sh, please name it as test_llms_text-generation_vllm-openvino.sh, I will let GHA trigger the test.

Signed-off-by: Zahidul Haque <zahidul.haque@intel.com>
@zahidulhaque
Copy link
Contributor Author

Please contribute an e2e test for this microservice, like this one: https://github.com/opea-project/GenAIComps/blob/main/tests/test_reranks.sh, please name it as test_llms_text-generation_vllm-openvino.sh, I will let GHA trigger the test.

@chensuyue, i am working on writing the test case and will create a separate PR for this.

@chensuyue chensuyue merged commit 7dbad07 into opea-project:main Jun 7, 2024
6 checks passed
@zahidulhaque zahidulhaque deleted the vllm-openvino-support branch June 7, 2024 05:52
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

9 participants