llama.cpp icon indicating copy to clipboard operation
llama.cpp copied to clipboard

LLM inference in C/C++

Results 1628 llama.cpp issues
Sort by recently updated
recently updated
newest added

### Name and Version version: 4726 (6dde1782) built with cc (GCC) 14.2.1 20250110 (Red Hat 14.2.1-7) for x86_64-redhat-linux ### Operating systems Linux ### GGML backends CPU ### Hardware AMD Ryzen...

bug-unconfirmed

This commit adds a script to build an XCFramework for iOS. The script uses CMake to build the library for both the simulator and device and then creates an XCFramework...

examples
devops

### Name and Version llama-cli ### Operating systems Linux ### GGML backends AMX ### Hardware XEON 8452Y + NV A40 ### Models _No response_ ### Problem description & steps to...

bug-unconfirmed

Using templates and reusing the `dequant_qX_Y` functions.

Nvidia GPU
ggml

### Name and Version convert_hf_to_gguf failed ### Operating systems _No response_ ### Which llama.cpp modules do you know to be affected? _No response_ ### Command line ```shell ``` ### Problem...

bug-unconfirmed

Originally, this work was part of #11515. Currently this macro doesn't work on files other than ggml-sycl.cpp because of a variable `g_ggml_sycl_debug` which gets initialized as a local variable only...

ggml
SYCL

Function `clip_build_img_from_pixels` is exported to build image from pixels. Application may load image with other libraries (e.g ffmpeg) for performance or more file formats.

examples

The commit updates the help view in the llama.swiftui example to use a NavigationView and a Done button to dismiss the help view. The motivation for this is that without...

examples

It's useful to be able to have this from the library layer as it's a key parameter of the model (e.g. to figure out how much KV cache memory is...