nunchaku
nunchaku copied to clipboard
[Bug] t5 encoder problem
i use it in flux comfyui workflow, the textencoder node seems needing three files,clip-L t5 and svdq-t5. The memory usage is huge,if we can only use clip-L.safetensor and svdq-t5.safetensors(3G file only),it will be great.
use int4 Loading checkpoint shards from combine these two model very slow
why just use svdq-t5.safetensors
Hi, we are optimizing the RAM usage for loading our T5 encoder. You can also use the FP8 text encoder by choosing the corresponding file and set the T5 precision to BF16 for now.
hope use t5 only svdq-t5.safetensors 2.99g
use dual clip gguf clip can run! so it is another way to save vram.
yeah, it is compatible.
hello, after use version 0.2,
use gguf clip loader easy to out of memory, clean vram and out of memory on the second queue
use nunchaku text encoder loader is ok, and plz support support only use 2.9g t5 model when use int4_model
#320 should fix this issue.
why are these files still needed? model-00001-of-00002.safetensors model-00002-of-00002.safetensors
fixed in v0.3.0
btw remember to use the v2 version node of nunchaku text encoder. 🤝