======
======
可以从zlibrary下载
I suggest you might consider using the [Vulkan version ](https://github.com/ggerganov/llama.cpp/blob/master/docs/build.md#vulkan)of llama.cpp, as it is compatible with almost all graphics cards.Intel's GPU software & driver is bad
I have fix some bug(half rope,mult-eos) in this pr https://github.com/ggml-org/llama.cpp/pull/12957 and use glm4 template as default
> > I have fix some bug(half rope,GGG output,mult-eos) in this pr [#12957](https://github.com/ggml-org/llama.cpp/pull/12957) and use glm4 template as default > > I quantized again GLM-9b-Z with the PR and it...
Using the Vulkan version seems to be more hassle-free.
Firstly, this is not a bug. In llama.cpp, it will prioritize checking for the presence of a dedicated graphics card. If none is found, it will then attempt to locate...
> > Firstly, this is not a bug. In llama.cpp, it will prioritize checking for the presence of a dedicated graphics card. If none is found, it will then attempt...
> > > > Firstly, this is not a bug. In llama.cpp, it will prioritize checking for the presence of a dedicated graphics card. If none is found, it will...