Forkoz

Results 474 comments of Forkoz

Will it load if you edit the model config and change "is_encoder_decoder": true, to false?

It's loading as the wrong type. This is a seq2seq.

I don't remember having issues with 1.23.5 and new TTS. I run XTTS and numpy keeps getting upgraded. I did upgrade librosa to the latest.

It's out.. you need a new DLL: https://github.com/TimDettmers/bitsandbytes/releases/tag/0.37.0

Which B*B you got installed.. and which PEFT.. I think the problem isn't in this repo.

Gotta make sure flash attention and torch version match.

Just tested and perf isn't good. 1.x It/s with no context. Maybe there is a way to just stop doing matmul in FP16? For reference this is how autoGPTQ does...

I've got 2 of them so if you need anything tested I can run it. I wish I was getting 160 t/s, but for some reason I'm not on the...