xidiancpy
xidiancpy
Have solve it?
add this code: size_array = [int(s) for s in x.size()[2:]] out = torch.nn.functional.avg_pool2d(x, size_array).view(batch, -1)
add upon the avg_pool2d 发自我的iPhone > 在 2020年11月7日,下午8:40,AlexanderGuthrie 写道: > > > add this code: > size_array = [int(s) for s in x.size()[2:]] > out = torch.nn.functional.avg_pool2d(x, size_array).view(batch, -1)...
sorry 发自我的iPhone > 在 2020年11月7日,下午10:38,AlexanderGuthrie 写道: > > > add upon the avg_pool2d 发自我的iPhone > … > 在 2020年11月7日,下午8:40,AlexanderGuthrie @.***> 写道: add this code: size_array = [int(s) for...
> [@tensorflowt](https://github.com/tensorflowt) Thanks for sharing! Currently, when the workload is too heavy (i.e., the decode cannot process all the requests in real time), there will be a known issue that...
@tensorflowt Could we add each other on WeChat to communicate together? If you agree, please send your WeChat ID to my email: [email protected]
> @chenqianfzh Thanks for your kindly help, this is very useful for me, and I have run successfully vLLM+LMCache DP disagg by your document, thanks so much! > > BTW,...
@maobaolong Could you please take a look at my mistakes for me?I set it up in accordance with the configuration mentioned above. Error streaming response from vLLM-2: Server error '500...
@chenqianfzh @maobaolong @rainj-me Could you please share your proxy.py file with me?
> > @chenqianfzh @maobaolong @rainj-me Could you please share your proxy.py file with me? > > https://github.com/bd-iaas-us/vllm/blob/lmcache_connector_from072/vllm/distributed/kv_transfer/kv_proxy/proxy.py > > Here you are @maobaolong Thank you for sharing. I have checked...