ipex-llm icon indicating copy to clipboard operation
ipex-llm copied to clipboard

Intel, are you OK?

Open yizhangliu opened this issue 1 month ago • 9 comments

We need you!

yizhangliu avatar Oct 31 '25 02:10 yizhangliu

Looks like they stopped developing IPEX-LLM...Good bye AI.

Dosperado74 avatar Nov 06 '25 16:11 Dosperado74

I guess that was the last Intel PC I bought for AI work...

erwinzierler avatar Nov 08 '25 22:11 erwinzierler

Are there any alternatives for LLM with Intel GPUs that is being maintained?

matheus2308 avatar Nov 13 '25 02:11 matheus2308

Have had success in Windows with the Vulkan version here: https://github.com/ggml-org/llama.cpp

ysaric avatar Nov 13 '25 04:11 ysaric

Also just saw that Vulkan can now be enabled with the offical Ollama. It worked with my GPU (Intel B50), although is a lot slower than ipex-llm (2-3x slower tokens/s), details here: https://github.com/ollama/ollama/releases/tag/v0.12.11

matheus2308 avatar Nov 14 '25 13:11 matheus2308

Also the new LM Studio Beta works with Vulkan and Intel GPUs: https://lmstudio.ai/beta-releases

ChrisPowers666 avatar Nov 15 '25 10:11 ChrisPowers666

Also just saw that Vulkan can now be enabled with the offical Ollama. It worked with my GPU (Intel B50), although is a lot slower than ipex-llm (2-3x slower tokens/s), details here: https://github.com/ollama/ollama/releases/tag/v0.12.11

On my Arc B580, running gemma3:12b it went from 35t/s with ipex to 15t/s with Vulkan.

charlescng avatar Nov 16 '25 19:11 charlescng

seems strange considering the B50 and B60...

hoiber avatar Nov 26 '25 23:11 hoiber

I have been playing with OpenVINO (https://github.com/openvinotoolkit/openvino) lately. It needs a bit more effort to get it running than LMstudio or Ollama but it really does make use of Intels CPU, GPU and NPU.

erwinzierler avatar Dec 06 '25 21:12 erwinzierler