FastChat
FastChat copied to clipboard
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Gpt2
## Why are these changes needed? Model worker (also vllm_worker) has error loading [Phi-3 models](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) "microsoft/Phi-3-mini-128k-instruct" etc..  ## Checks - [x] I've run `format.sh` to lint...
Hi! Thank you for your excellent work on the LLM evaluation! I'm inspired to create a French version of MT-Bench. Currently, I'm in the process of generating reference answers for...
Can we get a new release with llama3 support?
I would like to try 'top_k' in my scenario to see if anything different in the output, but seems the openai invoke method does not support this parameter. https://platform.openai.com/docs/api-reference/chat/create I...
I am running Fastchat on kubernetes. I have a worker for the controller, the fastchat api and a (gpu) worker for each of the models. When I pull this model...
What is the official system message for llama3 model ( `NousResearch/Meta-Llama-3-8B-Instruct`)? If I used the following in my code, however the system message is missing here: ``` register_conv_template( Conversation( name="llama-3",...
Please add this traditional Chinese model https://huggingface.co/taide
Hi, Willing to use the `Vicuna v1.5 7b` model for RAG or Retrieval Augmented Generation (Q&A based on retrieved documents or context), I tried many prompts based on Llama 2...
https://chat.lmsys.org/?leaderboard mentions `GPT-3.5-Turbo-0314` but I believe it is instead `GPT-3.5-Turbo-0301`, as indicated in `model_adapter.py`: https://github.com/lm-sys/FastChat/blob/f2e6ca964af7ad0585cadcf16ab98e57297e2133/fastchat/model/model_adapter.py#L71