ragflow icon indicating copy to clipboard operation
ragflow copied to clipboard

How do I call a model deployed using fastchat?

Open ciaoyizhen opened this issue 1 year ago • 1 comments

Describe your problem

Reading the related issue, it says to use ollama to start a local model, but https://ollama.com/library doesn't support ChatGLM,or needs a lot of work to support ChatGLM with ollama, also, currently already using fastchat to deploy other apps, so would like to be able to reuse this model, please Can I start a big model using fastchat and wrap the interface myself using fastapi, disguised as ollama? What are the key interfaces I need to provide to ragflow?

ciaoyizhen avatar Apr 23 '24 02:04 ciaoyizhen

相同需求,ollama就是个玩具儿,太难用了,统一按照OpenAI格式接入不就行了,已经成为业内规范了。ollama官方给的模型都是些4bit量化的,想加个自定义模型还要自己转换格式,也没有vllm推理优化。

Logistic98 avatar Jun 14 '24 03:06 Logistic98