Forkoz
Forkoz
Will it load if you edit the model config and change "is_encoder_decoder": true, to false?
It's loading as the wrong type. This is a seq2seq.
Consider editing the deps.
I don't remember having issues with 1.23.5 and new TTS. I run XTTS and numpy keeps getting upgraded. I did upgrade librosa to the latest.
It's out.. you need a new DLL: https://github.com/TimDettmers/bitsandbytes/releases/tag/0.37.0
Which B*B you got installed.. and which PEFT.. I think the problem isn't in this repo.
flash attention is broken
Gotta make sure flash attention and torch version match.
Just tested and perf isn't good. 1.x It/s with no context. Maybe there is a way to just stop doing matmul in FP16? For reference this is how autoGPTQ does...
I've got 2 of them so if you need anything tested I can run it. I wish I was getting 160 t/s, but for some reason I'm not on the...