transformers.js
transformers.js copied to clipboard
Is converting a Gemma 2B quantized compatible with transformers.js/onnx?
Question
I'm new to dev and wanted to know if converting a gemma 2b using the Optimum converter would work for this model?
The tokenizer is supported, but from a quick search in the source code it seems it's not itself supported. You can try WebLLM or Wllama to run that model? That's what I do.
@flatsiedatsie thx for the insight.
i'm thinking of converting the new qwen 2.5 0.5B model.
That should be supported, right?
update: looks like it's already supported! https://huggingface.co/onnx-community/Qwen2.5-0.5B-Instruct