hipudding

Results 63 comments of hipudding

> My question is: do I have to build on Huawei NPU server? Will the files I build in normal linux work on the Huawei NPU server? I don't know...

@AspartameJ 310p is not support by Ollama's inference engine(llama.cpp) currently.

There are symbolic links in the ascend-toolkit path, which causes ambiguity in relative paths. The symbolic links in libPath need to be removed before performing a Join operation. Please add...

> https://www.hikunpeng.ru/zh/developer/devkit/compiler/gcc I think the it's due to GCC version. Please use GCC 11.4 or higher version and try again.

@MeiK2333 It's due to llama.cpp inference engine. First, please confirm how many graphs are create when inferencing a model. If graphs counts are very big(say more than 100), it means...

@ruisearch42 @kevin85421 Could you please review this bug fix? Thanks.

Test case passwd: 2701/2701 tests passed Backend CANN0: OK Backend 2/3: CANN1 Skipping Backend 3/3: CPU Skipping 3/3 backends

> Additionally, I think we still need to verify that there are currently no memory leaks. I ran Valgrind on llama-server and didn’t find any leaks. There were a few...

> 在310P3上 编译报错 /llama.cpp/ggml/src/ggml-cann/aclnn_ops.cpp:2628:24: 错误:cannot convert ‘acl_tensor_ptr’ {aka ‘std::unique_ptr’} to ‘aclTensor*’ My bad. Forget to run testcase in 310p. see #17347

Need add a 310p compile check in CI. @xuedinge233