optimum
optimum copied to clipboard
似乎不支持cpu
只能支持gpu,但是官方gptq库的cpu已经支持了 if not torch.cuda.is_available(): raise RuntimeError("No GPU found. A GPU is needed to quantize model.")
Gently pinging @SunMarc and @fxmarty, do we have a plan on supporting gptq on CPUs?
Hi @andeyeluguo, you can't quantize a model using gptq quantization scheme without GPUs since it is not supported + it would take way too much time. However, you can run a gptq model with CPU if this is your use case ! Check the documentation here.