bloomz.cpp
bloomz.cpp copied to clipboard
Is there any plan for mT0 model support?
BLOOMZ and mT0 models are related, and mT0-13B performs better than BLOOMZ-176B in some cases.
The mT0-13B will be a killer model for normal user devices after a GPTQ-4bit quantization.
Hope the model can be supported.
Hopefully soon https://github.com/ggerganov/ggml/pull/12 👀