BUJIDAOVS
Results
4
comments of
BUJIDAOVS
trafficstars
一样,docker部署的72b-int4模型,单卡和双卡推理都非常慢
windows支持这么差吗,从安装到推理量化部署,报错没停过。有在win上跑通的案例吗
+1,希望能支持本地部署的模型,我的本地模型兼容支持openapi
> Thanks again for your dedicated contributions. > > As I was testing the functionality, how are we expected to use this feature? Currently, I use the following commands after...