Prince Canuma
Prince Canuma
My pleasure! Parallel calls is pretty much clear if we solve the first two.
I saw it, but it was not very informative on how to do it. I have my custom library for inferencing Models locally on Apple Silicon that follows the openAI...
Try changing the whisper model from `medium` to `tiny`. It worked for me :)
Thanks @rampadc, this is a great addition!
Please run pre-commit :)
@ivanfioravanti could use your help to double check this PR :)
Closing because #153 fixed
> I am having trouble segmenting multiple objects when using PaliGemma 2 mix ("mlx-community/paligemma2-3b-mix-448-bf16", "mlx-community/paligemma2-10b-mix-448-8bit"). I also tried to directly use transformers and with the 3B model I sometimes get...
If you could share the transformers examples as well would be nice Preferably with the images
Yes, the problem of models like this and some OCR models like DeepSeek-OCR is that prompt matters. And for such tasks it's best to use bf16 or fp16, quants struggle...