llama.cpp
llama.cpp copied to clipboard
LLM inference in C/C++
### Name and Version version: 4726 (6dde1782) built with cc (GCC) 14.2.1 20250110 (Red Hat 14.2.1-7) for x86_64-redhat-linux ### Operating systems Linux ### GGML backends CPU ### Hardware AMD Ryzen...
This commit adds a script to build an XCFramework for iOS. The script uses CMake to build the library for both the simulator and device and then creates an XCFramework...
### Name and Version llama-cli ### Operating systems Linux ### GGML backends AMX ### Hardware XEON 8452Y + NV A40 ### Models _No response_ ### Problem description & steps to...
Using templates and reusing the `dequant_qX_Y` functions.
### Name and Version convert_hf_to_gguf failed ### Operating systems _No response_ ### Which llama.cpp modules do you know to be affected? _No response_ ### Command line ```shell ``` ### Problem...
Originally, this work was part of #11515. Currently this macro doesn't work on files other than ggml-sycl.cpp because of a variable `g_ggml_sycl_debug` which gets initialized as a local variable only...
Function `clip_build_img_from_pixels` is exported to build image from pixels. Application may load image with other libraries (e.g ffmpeg) for performance or more file formats.
The commit updates the help view in the llama.swiftui example to use a NavigationView and a Done button to dismiss the help view. The motivation for this is that without...
It's useful to be able to have this from the library layer as it's a key parameter of the model (e.g. to figure out how much KV cache memory is...