ipex-llm icon indicating copy to clipboard operation
ipex-llm copied to clipboard

Running benchmark/all-in-one with GLM-4-9B-Chat model report "AutoTP not support for models"

Open dukelee111 opened this issue 1 year ago • 1 comments

Please help to confirm if the GLM-4-9B-Chat is supported , thanks so much.

Docker images:intelanalytics/ipex-llm-serving-vllm-xpu-experiment
Tag:2.1.0b2
Image ID:0e20af44ad46

step: cd /benchmark/all-in-one edit config.yaml bash run-deepspeed-arc.sh

Attached the error trace details: CHATGLM4-9B-Trace

dukelee111 avatar Aug 15 '24 00:08 dukelee111

Hi @dukelee111 ,

I reproduced and got the same error. "Not able to determine model policy automatically means that GLM-4-9B-Chat is not supported by AutoTP as shown here. It is not found in deepspeed's supported model list.

Uxito-Ada avatar Aug 19 '24 08:08 Uxito-Ada