Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Conversation

@tikikun
Copy link
Contributor

@tikikun tikikun commented Oct 11, 2023

Example curl to load model

curl -X POST 'http://localhost:3928/inferences/llamacpp/loadModel' \
-H 'Content-Type: application/json' \
-d '{
    "llama_model_path": "/Users/alandao/Documents/codes/nitro.cpp_temp/models/llama2_7b_chat_uncensored.Q4_0.gguf",
    "ctx_len": 2048,
    "ngl": 100,
    "embedding": true
}'

@tikikun tikikun self-assigned this Oct 11, 2023
@tikikun tikikun linked an issue Oct 11, 2023 that may be closed by this pull request
@tikikun tikikun merged commit ac5c5be into main Oct 11, 2023
@tikikun
Copy link
Contributor Author

tikikun commented Oct 11, 2023

Example curl to load model

curl -X POST 'http://localhost:3928/inferences/llamacpp/loadModel' \
-H 'Content-Type: application/json' \
-d '{
    "llama_model_path": "/Users/alandao/Documents/codes/nitro.cpp_temp/models/llama2_7b_chat_uncensored.Q4_0.gguf",
    "ctx_len": 2048,
    "ngl": 100,
    "embedding": true
}'

@hiro-v hiro-v deleted the 66-feat-loadunload-model-with-config-at-runtime branch October 13, 2023 23:27
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

feat: Load/unload model with config at runtime

2 participants