Jason Dai

Results 106 comments of Jason Dai

@MeouSker77 please take a look

It uses iGPU; as mentioned in readme, please refer to [[2]](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html)[[3]](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-microsoft-phi-3-models-intel-ai-soln.html)[[4]](https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-accelerate-alibaba-qwen2-llms.html) for more details.

> Hi @cyita , does that mean that [ollama with ipex-llm](https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/ollama_quickstart.md?rgh-link-date=2024-09-24T14%3A53%3A15Z) _does_ support flash attention for supported models automatically? > > As far as I know, the ipex-llm backend does...

> I'm in the same boat. I recently bought a NUC14 with a Intel Arc Graphics (iGPU 7Xe/112EU/896SP, Xe-LPG / Gen 12.7) but I'm unable to use Intel OneAPI because...

> Where is “ollama-ipex-llm-2.3.0b20250428-ubuntu.tgz“”? See https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/ollama_portable_zip_quickstart.md

> So slow update, sad. Intel , are you OK? Our current version is consistent with [v0.4.6](https://github.com/ollama/ollama/releases/tag/v0.4.6) of ollama. See https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/ollama_quickstart.md

@plusbang please take a look

Also update links in https://github.com/intel-analytics/ipex-llm/blob/main/README.zh-CN.md

> I've tried the official gemma3 models in 4b and 12b, as well as the q4_K_M versions from ollama, and then also `lmstudio-community/gemma-3-4b-it-GGUF:Q4_K_M` from huggingface and none of them seem...