-
Notifications
You must be signed in to change notification settings - Fork 13.5k
Closed
Labels
bugSomething isn't workingSomething isn't workinggood first issueGood for newcomersGood for newcomerslow severityUsed to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
Description
What happened?
-ctk, --cache-type-k TYPE KV cache data type for K (default: f16)
(env: LLAMA_ARG_CACHE_TYPE_K)
-ctv, --cache-type-v TYPE KV cache data type for V (default: f16)
(env: LLAMA_ARG_CACHE_TYPE_V)
It gives the default, but what are the other choices? I've been googling and can't find it anywhere.
Name and Version
docker run --rm --runtime nvidia --gpus all ghcr.io/ggerganov/llama.cpp:server-cuda --help
What operating system are you seeing the problem on?
Windows
Relevant log output
-ctk, --cache-type-k TYPE KV cache data type for K (default: f16)
(env: LLAMA_ARG_CACHE_TYPE_K)
-ctv, --cache-type-v TYPE KV cache data type for V (default: f16)
(env: LLAMA_ARG_CACHE_TYPE_V)Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't workinggood first issueGood for newcomersGood for newcomerslow severityUsed to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)