llama.cpp icon indicating copy to clipboard operation
llama.cpp copied to clipboard

readme : update gpt4all instructions

Open prusnak opened this issue 1 year ago • 14 comments

fixes https://github.com/ggerganov/llama.cpp/issues/975

prusnak avatar Apr 14 '23 19:04 prusnak

I am running into one issue with the bin:

python convert.py models/gpt4all-lora-quantized.bin
Loading model file models/gpt4all-lora-quantized.bin
Traceback (most recent call last):
  File "convert.py", line 1145, in <module>
    main()
  File "convert.py", line 1133, in main
    vocab = load_vocab(vocab_dir)
  File "convert.py", line 1075, in load_vocab
    raise FileNotFoundError(f"Could not find tokenizer.model in {path} or its parent; if it's in another directory, pass the directory as --vocab-dir")
FileNotFoundError: Could not find tokenizer.model in models or its parent; if it's in another directory, pass the directory as --vocab-dir```

robbintt avatar Apr 14 '23 19:04 robbintt

As the error says, you don't have the models/tokenizer.model file. You need to obtain it from the same place you are obtaining your LLaMa weights.

prusnak avatar Apr 14 '23 19:04 prusnak

I am running into one issue with the bin:

Also, based off the model and message, you probably should be using the migrate script, not the convert script. That model file is already ggml, it just needs to be migrated to current version.

MillionthOdin16 avatar Apr 14 '23 19:04 MillionthOdin16

migrate script, not the convert script

there is no migrate script anymore, just one complex one-does-it-all convert script (https://github.com/ggerganov/llama.cpp/pull/545)

prusnak avatar Apr 14 '23 19:04 prusnak

There's not a migrate-ggml-2023-03-30-pr613.py script?

MillionthOdin16 avatar Apr 14 '23 19:04 MillionthOdin16

There's not a migrate-ggml-2023-03-30-pr613.py script?

No.

prusnak avatar Apr 14 '23 19:04 prusnak

Welp. Good to know. Unfortunate that it has the same exact name as a previous convert.py script that does a subset of tasks lol

MillionthOdin16 avatar Apr 14 '23 19:04 MillionthOdin16

Anyways this PR is just for the readme. I have another error, will share in 20 min.

On Fri, Apr 14, 2023 at 12:32 MillionthOdin16 @.***> wrote:

Welp. Good to know. Unfortunate that it has the same exact name as a previous convert.py script that does a subset of tasks lol

— Reply to this email directly, view it on GitHub https://github.com/ggerganov/llama.cpp/pull/980#issuecomment-1509123846, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAOZDW2DGXIYDFQYDV4WFUDXBGQ4HANCNFSM6AAAAAAW6Z2L5Y . You are receiving this because you commented.Message ID: <ggerganov/llama. @.***>

robbintt avatar Apr 14 '23 19:04 robbintt

Anyways this PR is just for the readme. I have another error, will share in 20 min.

Exactly, this PR is about fixing readme. Any subsequent errors are out of scope of this PR and should be fixed by future PRs.

prusnak avatar Apr 14 '23 19:04 prusnak

Since this section is about Gpt4all, it might be helpful to enumerate the required files

robbintt avatar Apr 14 '23 22:04 robbintt

Thanks for cleaning up my oversight.

comex avatar Apr 15 '23 02:04 comex

Converting to draft since it seems that gpt4all conversion is broken, see https://github.com/ggerganov/llama.cpp/issues/984

prusnak avatar Apr 15 '23 08:04 prusnak

Converting to draft since it seems that gpt4all conversion is broken, see #984

Fix for this error is in https://github.com/ggerganov/llama.cpp/pull/1031

GPT4All-J is not supported at the moment, which is reflected in the README

prusnak avatar Apr 17 '23 17:04 prusnak

Fixes https://github.com/ggerganov/llama.cpp/issues/1036

prusnak avatar Apr 18 '23 15:04 prusnak

suggest adding the SHASUMS so people can validate their conversions

hughobrien avatar Apr 22 '23 23:04 hughobrien