FastChat icon indicating copy to clipboard operation
FastChat copied to clipboard

Use fastchat with download vicuna cpp model

Open rohezal opened this issue 1 year ago • 2 comments

Vicuna was build and worked as main.cpp by mself. But after this, I installed fastchat and it can't use the vicuna.cpp model file. Is it possible to fix this?

(fastchat) rohezal@box:/media/ai/fastchat/FastChat$ python3 -m fastchat.serve.cli --model-path /media/ai/llama/llama.cpp/models/ --device cpu
│                                                                                                  │
│ /home/rohezal/anaconda3/envs/fastchat/lib/python3.10/site-packages/transformers/utils/hub.py:380 │
│ in cached_file                                                                                   │
│                                                                                                  │
│    377 │   │   resolved_file = os.path.join(os.path.join(path_or_repo_id, subfolder), filename)  │
│    378 │   │   if not os.path.isfile(resolved_file):                                             │
│    379 │   │   │   if _raise_exceptions_for_missing_entries:                                     │
│ ❱  380 │   │   │   │   raise EnvironmentError(                                                   │
│    381 │   │   │   │   │   f"{path_or_repo_id} does not appear to have a file named {full_filen  │
│    382 │   │   │   │   │   f"'https://huggingface.co/{path_or_repo_id}/{revision}' for availabl  │
│    383 │   │   │   │   )                                                                         │
╰──────────────────────────────────────────────────────────────────────────────────────────────────╯
OSError: /media/ai/llama/llama.cpp/models/ does not appear to have a file named config.json. Checkout 'https://huggingface.co//media//ai/llama/llama.cpp/models//None' for 
available files.
(fastchat) rohezal@box:/media/ai/fastchat/FastChat$ python3 -m fastchat.serve.cli --model-path /media/ai/llama/llama.cpp/models/ggml-vic13b-uncensored-q5_1.bin --device cpu

│                                                                                                  │
│ /home/rohezal/anaconda3/envs/fastchat/lib/python3.10/site-packages/huggingface_hub/utils/_valida │
│ tors.py:160 in validate_repo_id                                                                  │
│                                                                                                  │
│   157 │   │   raise HFValidationError(f"Repo id must be a string, not {type(repo_id)}: '{repo_   │
│   158 │                                                                                          │
│   159 │   if repo_id.count("/") > 1:                                                             │
│ ❱ 160 │   │   raise HFValidationError(                                                           │
│   161 │   │   │   "Repo id must be in the form 'repo_name' or 'namespace/repo_name':"            │
│   162 │   │   │   f" '{repo_id}'. Use `repo_type` argument if needed."                           │
│   163 │   │   )                                                                                  │
╰──────────────────────────────────────────────────────────────────────────────────────────────────╯
HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/media//ai/llama/llama.cpp/models/ggml-vic13b-uncensored-q5_1.bin'. Use `repo_type` argument if needed.

rohezal avatar May 25 '23 18:05 rohezal

faced the same issue

kingwap99 avatar Jun 13 '23 06:06 kingwap99

I am now just using ooga booga:

https://github.com/oobabooga/text-generation-webui

rohezal avatar Jun 13 '23 07:06 rohezal