markluofd
markluofd
PaddleLite v2.11 Host环境: Ubuntu16.04 交叉环境 build_linux.sh配置: ARCH=armv7 WITH_PYTHON=ON PY_VERSION="3.5" 最后报错信息如下:  想请问交叉编译 with_python=ON,这个报错应该怎么处理?
### Proposal to improve performance _No response_ ### Report of performance regression _No response_ ### Misc discussion on performance I am using vllm to deploy the qwen 7b chat model...
根据文档里面提示,确认我手上的rk3399pro开发板是pcie模式,全局搜索npu_fw目录,发现在/usr/share下面有三个类似目录,分别是 npu_fw、npu_fw_pcie、npu_fw_pcie_optimization,这种情况应该更新哪个目录下的固件? 备注: 这边是一年前刷的rk3399pro官方的ubuntu18.04的操作系统
请问当前rk3399pro 1.7.3版本的driver和c api是稳定的吗?准备集成一下哈哈
### Describe the bug with demo run_llama_int8.py, setting generate_kwargs["do_sample"] to be True, I got the error as follows: command: python run_llama_int8.py -m ${MODEL_ID} --quantized-model-path "/workspace/saved_results/best_model.pt" --benchmark --jit --int8-bf16-mixed --num-iter 5...
2 H800 node deploying DeepSeek R1 model Using k8s deployment. If the pod only opens the server port and the port agreed by dist-init-addr, the worker will report an error...