gpt4all
gpt4all copied to clipboard
Python binding logs console errors when CUDA is not found, even when CPU is requested
Bug Report
Whichever Python script I run, when calling the GPT4All() constructor, say like this:
model = GPT4All(model_name='openchat-3.6-8b-20240522-Q5_K_M.gguf', allow_download=False, device='cpu')
...I get the following error messages:
Failed to load llamamodel-mainline-cuda-avxonly.dll: LoadLibraryExW failed with error 0x7e
Failed to load llamamodel-mainline-cuda.dll: LoadLibraryExW failed with error 0x7e
After that, the script continues to run normally, but these spurious error messages are annoying, particularly since:
- I don't have CUDA stuff installed
- ... because I don't have an nVidia GPU
- ... and I'm explicitly using CPU inference anyway
This did not happen in the earlier versions. I suspect it might be due to upstream changes in llama.cpp, but I'm not sure.
Your Environment
- Bindings version: 2.7.0
- Operating System: Windows 10
- Chat model used (if applicable): n/a