-
Notifications
You must be signed in to change notification settings - Fork 7.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Crash due to unhandled exception from ggml_vk_allocate in llama_kv_cache_init #1870
Comments
It'd be helpful if you could build GPT4All from source as described here with your build configuration set to Debug, open the directory containing the built |
@cebtenzzre here is the output of Please let me know if anything else is required |
Could you please run It would also be useful to have the output of |
This comment was marked as outdated.
This comment was marked as outdated.
Here is the output of |
As you can see, my system has both an integrated radeon GPU and a dedicated Nvidia one. I have tried gpt4all chat with all these system combinations on linux :
and the issue can be reproduced in all cases |
This comment was marked as outdated.
This comment was marked as outdated.
Sorry, that should be Which GPU do you have selected in the UI? |
It is showing
|
It was RTX 4060 I tried with CPU now and Wizard 1.2 is working. Just noticed that on Windows too its working via CPU instead of RTX 4060 with 8 GB VRAM My apologies for wasting your time. Should I close the issue? |
We still need to at least fix the fallback to CPU on Linux. Could you join the Discord and ping me? It seems like you're not getting the debug symbols that you should be - it may be more straightforward to just build with cmake instead. |
Sure, just did. |
Marking as fixed in the next release because of 6db5307 - testing would be appreciated. It seems like this can be reproduced easily by setting n_ctx to something really high - Mistral models will let you do this on the latest main, many others are limited to 4096 now. |
Fixed in v2.6.2. |
System Info
GPT4All version : 2.6.1
karthik@fedora:~$ free total used free shared buff/cache available Mem: 15633228 2679428 4617172 38508 8723496 12953800 Swap: 52236280 683776 51552504
karthik@fedora:~$ swapon NAME TYPE SIZE USED PRIO /dev/nvme0n1p6 partition 20G 0B 1 /dev/zram0 partition 29.8G 666.8M 100
Information
Reproduction
Expected behavior
The text was updated successfully, but these errors were encountered: