ollama is a model server aiding devs to run opensource models.
Open WebUI is an extensible, feature-rich, and user-friendly self-hosted WebUI designed to operate entirely offline. It supports various LLM runners, including Ollama and OpenAI-compatible APIs.
Shell Pilot is a version of the chatGPT-shell-cli library , modified to support Ollama and work with local LLM, and improve some features.
This repo contains assets that containerizes ollama, OpenWebUI and help us to run the whole stack either locally using Podman Compose or deploy them on OpenShift.
For Other Kuberenetes flavors, devs can replace route.yaml
with appropriate ingress.yaml
Running the stack locally:
git clone
this repo- cd into the project directory.
- Run
podman compose up
- Navigate to
https://localhost:8080
to access the GUI
- Review the values.yaml and update the required config
- Login into OpenShift using
oc login
if not logged in already - Create a new PVC called
model-store
with minimum size of 10 Gi - Run
helm upgrade --install my-chatgpt helm-charts/. -n <namespace>
- Once all the pods are up, navigate to the url as shown the
oc route
update the below env variables defined in compose.yml
or values.yaml
as required:
ENV Variable | Description | Default Value |
---|---|---|
PULL_MODEL_BY_DEFAULT | Determines if the model should be pulled automatically. | false |
MODEL | Name of the model to be pulled if auto-pulling is enabled. | llama3 |