gpt4all icon indicating copy to clipboard operation
gpt4all copied to clipboard

Python binding logs console errors when CUDA is not found, even when CPU is requested

Open brankoradovanovic-mcom opened this issue 7 months ago • 8 comments

Bug Report

Whichever Python script I run, when calling the GPT4All() constructor, say like this:

model = GPT4All(model_name='openchat-3.6-8b-20240522-Q5_K_M.gguf', allow_download=False, device='cpu')

...I get the following error messages:

Failed to load llamamodel-mainline-cuda-avxonly.dll: LoadLibraryExW failed with error 0x7e
Failed to load llamamodel-mainline-cuda.dll: LoadLibraryExW failed with error 0x7e

After that, the script continues to run normally, but these spurious error messages are annoying, particularly since:

  1. I don't have CUDA stuff installed
  2. ... because I don't have an nVidia GPU
  3. ... and I'm explicitly using CPU inference anyway

This did not happen in the earlier versions. I suspect it might be due to upstream changes in llama.cpp, but I'm not sure.

Your Environment

  • Bindings version: 2.7.0
  • Operating System: Windows 10
  • Chat model used (if applicable): n/a

brankoradovanovic-mcom avatar Jul 02 '24 18:07 brankoradovanovic-mcom