ipex-llm
ipex-llm copied to clipboard
Run neural-chat 7b inference with Deepspeed on Flex 140. #10507
Hi,
After review the previous issue : https://github.com/intel-analytics/ipex-llm/issues/10507.
We tested on Flex140 same suggestion, we get the performance very slow on Flex140 with both GPU running.
The xpu-smi GPU just usage very low. Suppose is 12GB/2 each GPU is 6GB.
Attached the platform setup configuration, see any need to addon.