BbBbB

Results 2 comments of BbBbB

I just ran llama-cli -m F:\models\ollama\gemma-3-12b-it-q4_0.gguf -p "Hi" --verbose ... load_tensors: tensor 'token_embd.weight' (f16) (and 0 others) cannot be used with preferred buffer type CPU_AARCH64, using CPU instead ... llama_kv_cache_init:...

> I have seen a similar error several times while running `qwen2.5-coder:32b-instruct-q5_K_M`. > > `ollama version is 0.5.4-ipexllm-20250228` ^^^^^^^^^^^^^^^^^^^^^^^^^^^ wonder how to check this version? I run ollama --version, it...