gpt4all icon indicating copy to clipboard operation
gpt4all copied to clipboard

[Feature] GPU support for GPT-J

Open 3Simplex opened this issue 1 year ago • 2 comments

Bug Report

When using Mac set to use metal, gpt-j model fails to fallback to CPU.

Steps to Reproduce

  1. With a Mac set application device to use metal.
  2. Try to load a gpt-j architecture model.
  3. With a Mac set application device to use CPU.
  4. Load the same gpt-j architecture model.

Expected Behavior

The gpt-j model has no GPU support should fallback to CPU.

Your Environment

  • GPT4All version: 2.8.0
  • Operating System: Mac
  • Chat model used (if applicable): https://huggingface.co/EleutherAI/gpt-j-6b converted to gguf.

Works in windows as expected with Vulkan.

3Simplex avatar Jun 12 '24 23:06 3Simplex

From Discord: qui — Today at 7:13 PM Yes. Happened when set to auto as well - in MAC Also failed when set to CUDA on Windows.

3Simplex avatar Jun 12 '24 23:06 3Simplex

GPU support for GPT-J would basically require upstreaming it into llama.cpp. It's the one architecture that we still support using an old implementation from the original ggml repo.

cebtenzzre avatar Jun 13 '24 14:06 cebtenzzre

Support for GPT-J models was removed in #2676. If someone makes a proper llama.cpp implementation of GPT-J, it should get GPU support automatically.

cebtenzzre avatar Aug 04 '24 21:08 cebtenzzre