Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

v2.7.3 crashes when loading large models, where v2.5.1 did not #2182

Open
dailysoftware opened this issue Mar 30, 2024 · 6 comments
Open

v2.7.3 crashes when loading large models, where v2.5.1 did not #2182

dailysoftware opened this issue Mar 30, 2024 · 6 comments
Labels
bug-unconfirmed chat gpt4all-chat issues need-info Further information from issue author is requested

Comments

@dailysoftware
Copy link

Bug Report

GPT4ALL crashes without any warning when using a model with RAM requirements greater than 16 GB. But when I switch version to 2.5.1 or loading a model with RAM requirements under 8GB, there is no problem.

Steps to Reproduce

1.Run GPT4ALL
2.Choose model
3.GPT4All then crashes

Your Environment

  • GPT4All version:2.7.3
  • Operating System:Win11
  • Chat model used (if applicable):nous-hermes-llama2-13b.Q4_0.gguf
@dailysoftware dailysoftware added bug-unconfirmed chat gpt4all-chat issues labels Mar 30, 2024
@dailysoftware
Copy link
Author

2.5.1 version can use various models, but the device can only use CPU and not GPU, with an error message indicating that GPU loading failed (out of VRAM). On the other hand, 2.7.3 version cannot use memory requirements of 16GB models, but it can use 8GB models and also use GPUs

@cebtenzzre
Copy link
Member

How much RAM do you have? Do you think it is possible that GPT4All is running out of RAM (e.g. does it crash when you set the device to "CPU"), or is it really crashing when it runs out of VRAM? The latter is possible, but it would definitely be a bug and not an intentional occurrence.

@cebtenzzre cebtenzzre reopened this Apr 1, 2024
@cebtenzzre cebtenzzre added the need-info Further information from issue author is requested label Apr 1, 2024
@cebtenzzre cebtenzzre changed the title Gpt4All crashes when loading models v2.7.3 crashes when loading large models, where v2.5.1 did not Apr 1, 2024
@Syclusion
Copy link

I am having this issue as well, 4090 and 96gb of memory. Running on cpu fixes crash but runs slow af

@TREHAND-Christian
Copy link

TREHAND-Christian commented Apr 22, 2024

I have the same problem, 80GB memory, NVIDIA RTX 3060.

QML debugging is enabled. Only use this in a safe environment.
[Debug] (Mon Apr 22 06:20:54 2024): deserializing chat "F:/AI/gpt4all/nomic.ai/GPT4All//gpt4all-3ca3afb4-8c17-4c97-8693-135477a84612.chat"
[Debug] (Mon Apr 22 06:20:54 2024): deserializing chats took: 4 ms
llama_new_context_with_model: max tensor size = 102.54 MB
llama.cpp: using Vulkan on NVIDIA GeForce RTX 3060
error loading model: Memory type index for buffer creation not found
llama_load_model_from_file_internal: failed to load model
LLAMA ERROR: failed to load model from F:/AI/gpt4all/nomic.ai/GPT4All/wizardcoder-python-34b-v1.0.Q4_0.gguf
GGML_ASSERT: C:\msys64\home\Jared\gpt4all-navarro\gpt4all-backend\llama.cpp-mainline\llama.cpp:552: data

@ItsCheif
Copy link

ItsCheif commented Jun 4, 2024

Issue seems to still exist on v2.8.0.

I've just got a large model that crashes GPT4ALL without warning, switched to CPU and it doesn't crash anymore. But it also just takes forever to write a single letter.

@securityopa
Copy link

I have the same problem with the latest version from flathub.

I have 128GB of RAM and I am using AMD Radeon 6800XT which is pretty fast in generating answers. But suddenly when the response is large, it crashes.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug-unconfirmed chat gpt4all-chat issues need-info Further information from issue author is requested
Projects
None yet
Development

No branches or pull requests

6 participants