[Feature] GPU support for GPT-J
Bug Report
When using Mac set to use metal, gpt-j model fails to fallback to CPU.
Steps to Reproduce
- With a Mac set application device to use metal.
- Try to load a gpt-j architecture model.
- With a Mac set application device to use CPU.
- Load the same gpt-j architecture model.
Expected Behavior
The gpt-j model has no GPU support should fallback to CPU.
Your Environment
- GPT4All version: 2.8.0
- Operating System: Mac
- Chat model used (if applicable):
https://huggingface.co/EleutherAI/gpt-j-6bconverted to gguf.
Works in windows as expected with Vulkan.
From Discord: qui — Today at 7:13 PM Yes. Happened when set to auto as well - in MAC Also failed when set to CUDA on Windows.
GPU support for GPT-J would basically require upstreaming it into llama.cpp. It's the one architecture that we still support using an old implementation from the original ggml repo.
Support for GPT-J models was removed in #2676. If someone makes a proper llama.cpp implementation of GPT-J, it should get GPU support automatically.