Jiaxin Shan
Jiaxin Shan
@ying2025 Did you check whether all your pods are ready? Seems there's only 1 visible GPU.
I moved to v0.1.0 since potential discussion and changes may needed
Move to v0.3.0 due to limited resources
> Support multiple models in the same server Does vLLM need the multi-model support similar like what FastChat does or something else?
@gangmuk did you check the logs? BTW, why not update your yaml to use v0.2.0-rc.2 for the runtime? I think you forgot to upgrade the other image 
> What log do you mean? pod log? Yes, pod logs. In that case, we know more details which container is not up, what's the reason it's not up. >...
Please describe podautoscaler next time, we know more details like the condition and events and more controller-manager-logs. ``` kubectl describe podautoscaler llama2-70b-pa ``` Does this issue still exist? Seems this...
@gangmuk I remember you did some follow up autoscaling testing. is this issue still not resolved? We supposed to make it work at least around March. I notice this issue...
@gangmuk did your first commit trigger the test? Check jobs trigger condition - https://github.com/vllm-project/aibrix/blob/main/.github/workflows/lint-and-tests.yml - https://github.com/vllm-project/aibrix/blob/main/.github/workflows/installation-tests.yml I feel the benchmark folder may not trigger any test at this moment