-
Notifications
You must be signed in to change notification settings - Fork 13.7k
Open
Labels
Description
Name and Version
$ ./llama-server --version
ggml_vulkan: Found 1 Vulkan devices:
ggml_vulkan: 0 = Intel(R) Iris(R) Xe Graphics (DG1) (Intel open-source Mesa driver) | uma: 0 | fp16: 1 | bf16: 0 | warp size: 32 | shared memory: 65536 | int dot: 1 | matrix cores: none
version: 7079 (416e7c7)
built with cc (Debian 14.3.0-8) 14.3.0 for x86_64-linux-gnu
Operating systems
Linux
Which llama.cpp modules do you know to be affected?
llama-server
Command line
./llama-server -m /path/to/models/ibm-granite_granite-4.0-h-micro-Q4_0.gguf --host HOSTNAME --port PORT -c 16384 -b 8192 -e -ngl 99 -t 8 -n -1 --no-mmap --jinja --kv-unified -fa off -np 2Problem description & steps to reproduce
I git pull'ed this morning then it went crazy on 416e7c7 So I rolled back all those vulkan related commits one by one until the one right before 38eaf32, it became ok. Then I fast forward'ed to 416e7c7 again and simply git revert 38eaf32, it was still ok. So I think 38eaf32 is the issue?
First Bad Commit
Relevant log output
(test-backend-ops -o MUL_MAT,MUL_MAT_ID)
4/4 tests passed
Backend BLAS: OK
Backend 3/3: CPU
Skipping CPU backend
3/3 backends passed
OK
(model loading and running)
Nothing unusual except output is gibberish