intel-extension-for-transformers
intel-extension-for-transformers copied to clipboard
catch prepack error and fallback to torch bf16
Type of Change
bug fix
Description
a solution to discussion https://github.com/intel/intel-extension-for-transformers/discussions/1506
Expected Behavior & Potential Risk
prepack error skipped
How has this PR been tested?
user's test
Dependency Change?
None
⚡ Required checks status: All passing 🟢
Groups summary
🟢 Format Scan Tests workflow
| Check ID | Status | Error details | |
|---|---|---|---|
| format-scan (pylint) | success | ✅ | |
| format-scan (bandit) | success | ✅ | |
| format-scan (cloc) | success | ✅ | |
| format-scan (cpplint) | success | ✅ |
These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py, intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/retrieval_agent.py.
🟢 NeuralChat Unit Test
| Check ID | Status | Error details | |
|---|---|---|---|
| neuralchat-unit-test-baseline | success | ✅ | |
| neuralchat-unit-test-PR-test | success | ✅ | |
| Generate-NeuralChat-Report | success | ✅ |
These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py, intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/retrieval_agent.py.
🟢 Chat Bot Test workflow
| Check ID | Status | Error details | |
|---|---|---|---|
| call-inference-llama-2-7b-chat-hf / inference test | success | ✅ | |
| call-inference-mpt-7b-chat / inference test | success | ✅ |
These checks are required after the changes to intel_extension_for_transformers/neural_chat/models/model_utils.py, intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/retrieval_agent.py.
Thank you for your contribution! 💜
Note This comment is automatically generated and will be updates every 180 seconds within the next 6 hours. If you have any other questions, contact VincyZhang or XuehaoSun for help.