O.T
O.T
pretty please
@dje-dev any updates?
@wilson1yan Can you share the shell/bash script for setting up the inference server via vLLM for PyTorch model, FP16? > If using vLLM for inference (PyTorch model, FP16), I believe...
Also just commenting to prevent closure of the issue since it is one that I am also tracking!
Restarted, still dont see anything ... (on windows)