intel-extension-for-transformers
intel-extension-for-transformers copied to clipboard
Llava Models
Does the Llava part work ?
https://github.com/intel/intel-extension-for-transformers/tree/main/intel_extension_for_transformers/transformers/modeling/llava_models
If so are they optimized for Intel Device and are there any examples ?
Thanks for building this library. I have seen the token generation performance to be very good compared to OpenVino.
Great work !
Thanks
Thanks for your usage!!! Unfortunately, we are not support LLaVA for now. But we may consider to support it in future, please stay tune.
Regrads, Bo
It's for multi-model training, but optimization is WIP.
@kevinintel how do you do the optimization for the llava model and use it ?
Someone tried low-bits for llava: https://arxiv.org/pdf/2306.00978.pdf and we will try to quantize it.
Thanks @kevinintel
Hi, support for quantization of multimodal models is currently planned, and any updates will be communicated here.
we can optimize llava in https://github.com/intel/neural-compressor/pull/1797 will add examples