Sharing my vLLM Docker Container Image #1454
samos123
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
Can you explain how I can interact with the docker container with Langchain for instance? Or can I use the HF text-inference framework and call this as an API endpoint? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Repo: https://github.com/substratusai/vllm-docker
This container image runs the OpenAI API server of vLLM.
Image URL:
ghcr.io/substratusai/vllm
Quickstart
Deploy Mistral 7B Instruct:
Configuration Options
The following configuration options are available by using environment
variables:
hf_org/model
or utilize a path to point to a local model. Example value: mistralai/Mistral-7B-Instruct-v0.1The container image automatically detects the number of GPUs and sets
--tensor-parallel-size
to be equal to number of GPUs available. Thegpu-count.py
script is used to detect number of GPUs.Building
Beta Was this translation helpful? Give feedback.
All reactions