FastChat
FastChat copied to clipboard
How to use multiple Ascend NPUs?
Why is the --device npu parameter fixed to support only one Ascend NPU in code instead of multiple NPUs?
def generate_stream_gate(self, params):
if self.device == "npu":
import torch_npu
torch_npu.npu.set_device("npu:0")
Same issue. Has this issue be solved?
I have the same issue, and hardcoding to multiple NPUs won't work either.
同样的问题
ASCEND_RT_VISIBLE_DEVICES=1
![]()
ASCEND_RT_VISIBLE_DEVICES=1
This only changes that one NPU being used, not using multiple NPUs.
请问问题解决了吗,我现在只能多卡加载模型,但是还是只能单卡推理