FastChat icon indicating copy to clipboard operation
FastChat copied to clipboard

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.

Results 766 FastChat issues
Sort by recently updated
recently updated
newest added

When running ``` python ../fastchat/eval/get_model_answer.py --model-id 1 \ --model-path ../vicuna \ --question-file t.jsonl \ --answer-file answer.jsonl \ --num-gpus 2 ``` I get the following error: ``` 2023-05-05 17:23:57,395 INFO worker.py:1616...

Currently we know from the dataset which model is better than another model. It would be useful to also have the different responses and their respective inputs, but I don't...

Hi, since I updated fastchat to version 0.2.2 I can no longer make the 4-bit GPTQ work because I get this error: `python3 -m fastchat.serve.cli --model-path models/TheBloke_vicuna-7B-1.1-GPTQ-4bit-128g --wbits 4 --groupsize...

Hi,, I have been trying to reproduce vicuna in 4xL4 which is 4*24=96 gb vram. What approach should I follow? Train.py oom error Train_mem.py oom error Left is Lora but...

https://huggingface.co/mosaicml/mpt-7b-storywriter The model appears to be a great alternative to LLaMA from surface observations, particularly for commercial use. I expect many groups to build upon this foundation going forward if...

good first issue

Hello, I've noticed that in the `FastChat/fastchat/conversation.py` file has code that interfaces with multiple LLMs. I think that as we continue to add more LLMs, there will come a time...

enhancement

Hi I noticed that Nvidia released the following model https://huggingface.co/nvidia/GPT-2B-001 Is this something we can run on the FastChat platform?

good first issue

Hi, Can I use GPTQ quantized model to do the inference? ``` https://github.com/qwopqwop200/GPTQ-for-LLaMa ``` Thank you!

hello, i can't load the model vicuna-13B-1.1-GPTQ-4bit-128g . i have installed git clone https://github.com/oobabooga/GPTQ-for-LLaMa.git -b cuda and python 3.9 thanks for your help. **python -m fastchat.serve.cli --model-name TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g --wbits 4...

How to fine tune vicuna-7b with A40

question