Liang Lv
Liang Lv
Hi @Akshaysharma29, Could you please try to convert your savedmodel to pb file and then use session run to check if your fp32 graph is good? You didn't evaluate the...
Hi @Akshaysharma29, There are built-in datasets 'bert' and 'mzbert' which are for the bert type of models. They use the tf_record format. The 'dummy' dataset is not suitable for bert...
@matthew-olson-intel The PR https://github.com/tensorflow/tensorflow/pull/53480 is related with Pad + Conv3D. It's not for Pad + Conv2D. Good news is that this issue has already been fixed by TF SPR-Base and...
@tbykowsk, Apologies for the delayed response regarding this issue. I hadn't realized it was assigned to me. You can refer to the [quick start example](https://github.com/intel/intel-extension-for-transformers/blob/main/intel_extension_for_transformers/neural_chat/examples/quick_start/chatbot/README.md) for setuping the chatbot service....
@bbelky, Please use python 3.10.
> Let's take a look how vLLM and TGI supports. we can leverage them and add what we need if missing. They are serving framework, so they don't support user...
Hi @olegmikul, To resolve the Chatbot issue, you'll need to install an additional requirements.txt file located at intel_extension_for_transformers/neural_chat/requirements_cpu.txt before running the chatbot. For the INT4 Inference issue, please execute `pip...