FastChat icon indicating copy to clipboard operation
FastChat copied to clipboard

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.

Results 766 FastChat issues
Sort by recently updated
recently updated
newest added

I have a peft adapter model for a finetuned Falcon7b model, When using gen_mode_answer.py, i get this error: TypeError: PeftModelForCausalLM.generate() takes 1 positional argument but 2 were given python gen_model_answer.py...

fschat==0.2.36 ``` WARNING 02-29 16:06:12 config.py:140] gptq quantization is not fully optimized yet. The speed can be slower than non-quantized models. INFO 02-29 16:06:12 llm_engine.py:72] Initializing an LLM engine with...

when I run ' python3 -m fastchat.serve.openai_api_server --host localhost --port 8000' then output error : Traceback (most recent call last): File "/work/miniconda3/envs/zq_vllm/lib/python3.10/runpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None,...

## Why are these changes needed? When using langchaingo w/ fastchat, it generates requests with formats like `"messages":["text":"given text", "type": "text"]` as default, which will be recognized and processed as...

Is FastChat not support M3E-BASE? it can run bge-m3 but can't run M3E-BASE

I used the local model about 40GB, then increased it to 50GB as the coversation increased, but couldn't free the GPU when exited the conversations

Hello,I used Qlora to train,but I get an error: NotImplementedError: Cannot copy out of meta tensor; no data! requriements.txt: ``` peft @ file:///root/peft torch==1.13.1+cu116 torchaudio==0.13.1+cu116 torchvision==0.14.1+cu116 transformers==4.28.1 deepspeed==0.9.4 flash-attn==0.2.0 ```...

## Why are these changes needed? Some models need to specifically learn to generate the stop tokens. Otherwise these trained models will not stop when serving. This is a model...

Hey. I haven't found similar examples for using FastChat for multi-node training. My script is as follows: I have found from this website https://pytorch.org/docs/stable/elastic/run.html to use the "rdzv-id, rdzv-backend and...

## Why are these changes needed? With this feature, we can use the guidance to controll fastchat's llm output. See this [issue2481](https://github.com/lm-sys/FastChat/issues/2481). ## Related issue number (if applicable) [issue2481](https://github.com/lm-sys/FastChat/issues/2481). ##...