ggml icon indicating copy to clipboard operation
ggml copied to clipboard

ggml vs Qualcomm SNPE inference engine on qualcomm soc

Open Francis235 opened this issue 9 months ago • 0 comments

Hello, I plan to deploy the model using ggml on Qualcomm's chip. I'm curious about the comparison between using ggml for inference on an SoC chip (such as a Qualcomm SoC, involving components like CPU, GPU, NPU, etc.) versus leveraging the inference engine provided by the chip itself (such as qualcomm SNPE). Since ggml inference primarily takes place on the CPU, whereas the chip's inference engine can offload computations to the GPU or NPU, does using ggml lead to a significant increase in CPU memory usage and %CPU, potentially impacting other tasks? Has anyone conducted a similar comparative test?

Francis235 avatar Apr 30 '24 02:04 Francis235