Awni Hannun
Awni Hannun
Yes that's been on our list to add for a while. Though, do you use `mlx_lm.lora` or just the original `lora.py` script? My preference is to update the package since...
Agree I think we may need to start using a yaml config
Yes indeed.. safe to close, thank you!
I actually just tried to use convert on the model and I got this issue `python -m mlx_lm.convert --hf-path bigcode/starcoder2-3b`: ``` File "/Users/awni/mlx-examples/llms/mlx_lm/utils.py", line 413, in fetch_from_hub config = AutoConfig.from_pretrained(model_path)...
Fix is here https://github.com/ml-explore/mlx-examples/pull/574
@danilopeixoto I've been thinking about having this in MLX LM recently. Any interest in sending a PR? It might make to do it after we have a more manageable config...
To be more concrete, I'm envisioning you just set the loss in the config. e.g. `cross_entropy` or `dpo`
Should be ready soon https://github.com/ml-explore/mlx/pull/809, although that will only run on the CPU so it may be too slow depending on how often you use it.
I'm not opposed to supporting bfloat but I would not want to make it the default: - float16 is still considerably faster given it has native support. The benchmarks in...
> I am wondering that there are any documents about how to accelerate SVD by GPU? I would start by learning about parallel implementations of SVD in general (maybe try...