stable-fast icon indicating copy to clipboard operation
stable-fast copied to clipboard

FP8 support in stable fast

Open jkrauss82 opened this issue 1 year ago • 6 comments

Is it planned?

Currently getting this error when trying to run ComfyUI in fp8 (flags --fp8_e4m3fn-text-enc --fp8_e4m3fn-unet):

RuntimeError: "addmm_cuda" not implemented for 'Float8_e4m3fn'

jkrauss82 avatar Feb 25 '24 15:02 jkrauss82

Is it planned?

Currently getting this error when trying to run ComfyUI in fp8 (flags --fp8_e4m3fn-text-enc --fp8_e4m3fn-unet):

RuntimeError: "addmm_cuda" not implemented for 'Float8_e4m3fn'

I'm quite sure stable fast has its own quantization node but it's not implemented in the node iirc

banjaminicc avatar Feb 25 '24 16:02 banjaminicc

@jkrauss82 Sorry, FP8 kernels aren't implemented and I guess I lack the time to support them now.

chengzeyi avatar Feb 26 '24 09:02 chengzeyi

Thanks for the reply, understood. It would be nice if it could be supported eventually.

jkrauss82 avatar Feb 27 '24 21:02 jkrauss82

@jkrauss82 I have created one new project which supports FP8 inference with diffusers. However, it has not been open-sourced. I hope it could be made publicly soon...

chengzeyi avatar May 09 '24 14:05 chengzeyi

Is it planned? Currently getting this error when trying to run ComfyUI in fp8 (flags --fp8_e4m3fn-text-enc --fp8_e4m3fn-unet):

RuntimeError: "addmm_cuda" not implemented for 'Float8_e4m3fn'

I'm quite sure stable fast has its own quantization node but it's not implemented in the node iirc

A new project could be published soon to support FP8 inference instead of stable-fast. I hope everyone will enjoy it.

chengzeyi avatar May 09 '24 15:05 chengzeyi

That would be very welcome. I have seen fp8 support is getting traction recently in the vllm project. Would be nice to have it in diffusers/img gen as well. I will stay tuned. Thanks for the update!

jkrauss82 avatar May 09 '24 21:05 jkrauss82