gpt4all icon indicating copy to clipboard operation
gpt4all copied to clipboard

v2.7.3 crashes when loading large models, where v2.5.1 did not

Open dailysoftware opened this issue 4 months ago • 4 comments

Bug Report

GPT4ALL crashes without any warning when using a model with RAM requirements greater than 16 GB. But when I switch version to 2.5.1 or loading a model with RAM requirements under 8GB, there is no problem.

Steps to Reproduce

1.Run GPT4ALL 2.Choose model 3.GPT4All then crashes

Your Environment

  • GPT4All version:2.7.3
  • Operating System:Win11
  • Chat model used (if applicable):nous-hermes-llama2-13b.Q4_0.gguf

dailysoftware avatar Mar 30 '24 16:03 dailysoftware

2.5.1 version can use various models, but the device can only use CPU and not GPU, with an error message indicating that GPU loading failed (out of VRAM). On the other hand, 2.7.3 version cannot use memory requirements of 16GB models, but it can use 8GB models and also use GPUs

dailysoftware avatar Mar 30 '24 16:03 dailysoftware

How much RAM do you have? Do you think it is possible that GPT4All is running out of RAM (e.g. does it crash when you set the device to "CPU"), or is it really crashing when it runs out of VRAM? The latter is possible, but it would definitely be a bug and not an intentional occurrence.

cebtenzzre avatar Apr 01 '24 19:04 cebtenzzre

I am having this issue as well, 4090 and 96gb of memory. Running on cpu fixes crash but runs slow af

Syclusion avatar Apr 15 '24 18:04 Syclusion

I have the same problem, 80GB memory, NVIDIA RTX 3060.  QML debugging is enabled. Only use this in a safe environment. [Debug] (Mon Apr 22 06:20:54 2024): deserializing chat "F:/AI/gpt4all/nomic.ai/GPT4All//gpt4all-3ca3afb4-8c17-4c97-8693-135477a84612.chat" [Debug] (Mon Apr 22 06:20:54 2024): deserializing chats took: 4 ms llama_new_context_with_model: max tensor size = 102.54 MB llama.cpp: using Vulkan on NVIDIA GeForce RTX 3060 error loading model: Memory type index for buffer creation not found llama_load_model_from_file_internal: failed to load model LLAMA ERROR: failed to load model from F:/AI/gpt4all/nomic.ai/GPT4All/wizardcoder-python-34b-v1.0.Q4_0.gguf GGML_ASSERT: C:\msys64\home\Jared\gpt4all-navarro\gpt4all-backend\llama.cpp-mainline\llama.cpp:552: data

TREHAND-Christian avatar Apr 22 '24 04:04 TREHAND-Christian