vicuna-installation-guide icon indicating copy to clipboard operation
vicuna-installation-guide copied to clipboard

how to generate a llama.cpp server with fastchat api

Open xx-zhang opened this issue 2 years ago • 1 comments
trafficstars

I have set the server , but only few words output like blocked , and is a single progress which can't reponse fastly. it is only run and load model when the request is getting.

xx-zhang avatar Jun 27 '23 01:06 xx-zhang

How did you setup the server?

fredi-python avatar Jun 27 '23 23:06 fredi-python