mlx-vlm
mlx-vlm copied to clipboard
Batch inference support (self-assigning)
Would be nice to have batch inference support similar to mlx_parallm, happy to try and add soon. @Blaizzy can you assign this to me?
Hey Will,
Yes, that would be awesome!
I have assigned the task to you 😀
Please comment #40 here so I can assign it to you and discuss details on a single issue.