Jiaxin Shan

Results 742 comments of Jiaxin Shan

@ying2025 Did you check whether all your pods are ready? Seems there's only 1 visible GPU.

I moved to v0.1.0 since potential discussion and changes may needed

> Support multiple models in the same server Does vLLM need the multi-model support similar like what FastChat does or something else?

@gangmuk did you check the logs? BTW, why not update your yaml to use v0.2.0-rc.2 for the runtime? I think you forgot to upgrade the other image ![Image](https://github.com/user-attachments/assets/9283a685-4d20-4272-8803-04517fa2932e)

> What log do you mean? pod log? Yes, pod logs. In that case, we know more details which container is not up, what's the reason it's not up. >...

Please describe podautoscaler next time, we know more details like the condition and events and more controller-manager-logs. ``` kubectl describe podautoscaler llama2-70b-pa ``` Does this issue still exist? Seems this...

@gangmuk I remember you did some follow up autoscaling testing. is this issue still not resolved? We supposed to make it work at least around March. I notice this issue...

@gangmuk did your first commit trigger the test? Check jobs trigger condition - https://github.com/vllm-project/aibrix/blob/main/.github/workflows/lint-and-tests.yml - https://github.com/vllm-project/aibrix/blob/main/.github/workflows/installation-tests.yml I feel the benchmark folder may not trigger any test at this moment