FastChat
FastChat copied to clipboard
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
…issue #2994) ## Why are these changes needed? 1. Add support for pydanticv2 and resolve incompatible with vllm 0.3.0 [issue #2994](https://github.com/lm-sys/FastChat/issues/2994). 2. Fix error when fine tuning is loading by...
Could you add chat template for CohereForAI/c4ai-command-r-plus ?
Whether this project is only adapted to the Ascend NPU 910B chip? I am trying to run fastchat vicuna-7b-v1.5 on Ascend NPU 910A chip, but the inference speed very very...
I was using `gpt-4`, and I found that `sep_style=None` in its template, so `get_prompt` got the error: ValueError: Invalid style: None 
Hi, sorry I'm a big newb at this. I love that fastchat uses multi gpus and I've gotten the webservice/worker setup, but I don't know how to config SillyTavern to...
I've noticed that the current implementation supports logprobs (thanks a lot!) but it appears to be different from how OpenAI handles them. I was wondering if it would be possible...
ChatGLM3-6B adopts a newly designed [Prompt format](https://github.com/THUDM/ChatGLM3/blob/main/PROMPT_en.md), supporting multi-turn dialogues as usual. It also natively supports [tool invocation](https://github.com/THUDM/ChatGLM3/blob/main/tools_using_demo/README_en.md) (Function Call), code execution (Code Interpreter), and Agent tasks in complex scenarios....
fastchat内部使用fastapi.是支持高并发吗?是否qwen1.5
# HuggingFace https://huggingface.co/Qwen/Qwen1.5-72B-Chat
Looking forward to adding support for Qwen1.5, including Qwen1.5-7B-Chat, Qwen1.5-7B-Chat-GPTQ-Int8, and so on. Qwen1.5 is more powerful than Qwen. Thank you.