HarryHang
Results
1
issues of
HarryHang
使用Model scope下载的 Qwen2.5-Omni-7B-MNN模型,在config.json里把后端改为cuda,在有两块4080的linux服务器上运行llm_demo,出现以下两个问题: 1、算子不支持 CUDABackend Don't support type Attention, Attention/Reshape_8_output_0 CUDABackend Don't support type Attention, Attention/Reshape_17_output_0 CUDABackend Don't support type Attention, Attention/Reshape_26_output_0 CUDABackend Don't support type Attention, Attention/Reshape_35_output_0 CUDABackend Don't...