======

Results 10 comments of ======

可以从zlibrary下载

I suggest you might consider using the [Vulkan version ](https://github.com/ggerganov/llama.cpp/blob/master/docs/build.md#vulkan)of llama.cpp, as it is compatible with almost all graphics cards.Intel's GPU software & driver is bad

I have fix some bug(half rope,mult-eos) in this pr https://github.com/ggml-org/llama.cpp/pull/12957 and use glm4 template as default

> > I have fix some bug(half rope,GGG output,mult-eos) in this pr [#12957](https://github.com/ggml-org/llama.cpp/pull/12957) and use glm4 template as default > > I quantized again GLM-9b-Z with the PR and it...

Using the Vulkan version seems to be more hassle-free.

Firstly, this is not a bug. In llama.cpp, it will prioritize checking for the presence of a dedicated graphics card. If none is found, it will then attempt to locate...

> > Firstly, this is not a bug. In llama.cpp, it will prioritize checking for the presence of a dedicated graphics card. If none is found, it will then attempt...

> > > > Firstly, this is not a bug. In llama.cpp, it will prioritize checking for the presence of a dedicated graphics card. If none is found, it will...