FastChat icon indicating copy to clipboard operation
FastChat copied to clipboard

fastchat.serve.openai_api_server doesn't work with `stream=true` parameter

Open richginsberg opened this issue 1 year ago • 2 comments
trafficstars

I'm trying to get FastChat working on https://github.com/open-webui/open-webui via it's option to accept OpenAI compatible endpoints. TabbyAPI seems to be able to serve OpenAI API format with streaming/chunking. FastChat doesn't seem to support.

I've install Exllamav2 and confirmed API works well

python3 -m fastchat.serve.controller --host 0.0.0.0 --port 21001

python3 -m fastchat.serve.model_worker --model-path ~/model/Llama-3-8B-Instruct-262k-8.0bpw-h6-exl2 --enable-exllama

python3 -m fastchat.serve.openai_api_server --host 0.0.0.0 --port 8000

Works fine w/o "stream": true no_stream_option

Issue

Errors with "stream": true stream_true

richginsberg avatar May 03 '24 02:05 richginsberg

I also encountered this problem, have you solved this?

nuocheng avatar May 18 '24 09:05 nuocheng

fastchat streams output tokens on another endpoint/module. Hoping it was in roadmap to port to fastchat.serve.openai_api_server

richginsberg avatar May 18 '24 10:05 richginsberg

Hi everyone! I have the same problem. Has anyone found a local solution?

kekslop avatar May 22 '24 10:05 kekslop