Pedro Cuenca

Results 331 comments of Pedro Cuenca

In terms of user expectations, if `--local-dir` is specified then it sounds reasonable to me that the global cache is not required. Totally agree with keeping the current behaviour in...

I just tested commit 0588d7cab707a8952c4433137ec70a49d864fe2b and the same error occurs. Trying to debug.

Hi @seboslaw! I've recently done a similar exercise, and discovered that if the model accepts flexible shapes, then Core ML only uses the CPU. In the case of sequence-to-sequence models...

> I've tried your suggestion (I think I did :) and updated the upperBounds of the input parameters Sorry, I think I wasn't clear. I didn't mean to make the...

Testing T5 is high up in my to-do list, I hope to get to it pretty soon and hopefully I'll have some insight then :) Sorry for the no-answer though.

@seboslaw What you tried to do [here](https://github.com/huggingface/exporters/issues/33#issuecomment-1547815946) used to work, but in newer versions of Core ML it results in the error you've seen. The problem is that the model...

@seboslaw This is the branch: https://github.com/huggingface/exporters/pull/37. I have other local changes, so I hope I didn't break or miss anything. I verified that T5 encoder and decoder export with fixed...

Hi @fakerybakery! I think the easiest way is to use this automated Space, which uses `exporters` under the hood: https://huggingface.co/spaces/huggingface-projects/transformers-to-coreml You enter the model id (`facebook/m2m100_418M`), then select the task...

> I can't convert 70b on this @USBhost did you try with `convert-hf-to-gguf.py`?

> Doesn't seem that the `eos_token` is working with either of the convert scripts in this PR @m18coppola the instruct models use two different EOS tokens: the standard one (``),...