transformers.js
transformers.js copied to clipboard
4bit ONNX models support
Feature request
It look like ONNX now supports 4bit: https://onnx.ai/onnx/technical/int4.html
It would be nice if we could use 4bit models with transformers.js .
Motivation
Make models for semantic-autocomplete that are even smaller when it comes to disk size
Your contribution
Maybe, but for the moment is not even clear to me how to convert models to 4bit. I think ONNX needs to add more docs.