Awni Hannun
Awni Hannun
I left a couple of high level questions, could you check? I will do a more detailed review after hearing from you. FYI usually with the examples reivewing for nits...
@menzHSE is this ready for review?
@menzHSE this is really nice! I took a deeper look today and simplifies / reorganized the example to be more self-contained and in the style of our other small examples....
@menzHSE you might be interested to check the updated README log. I notice the throughput I get on my M1 Max is much much better than the numbers you had...
Maybe you are using an outdated whisper? The `transcribe` function takes an `path_or_hf_repo` as the second argument which you should set to `mlx_models/large` if that's where you stored the model.
I think the README for this example needs a little work, I will update it.
For future reference `help(whisper.transcribe)` will tell you about the arguments it can accept 😄
Could you share the training command you used?
I think the problem here is that our `lora.py` script assumes you are sending it an "unfused" model so it is loading the adapters into a fused model and hence...
Fix in #282, thanks!