text-generation-inference
text-generation-inference copied to clipboard
add intel xpu support for TGI
What does this PR do?
Fixes # (issue)
Before submitting
- [ ] This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
- [ ] Did you read the contributor guideline, Pull Request section?
- [ ] Was this discussed/approved via a Github issue or the forum? Please add a link to it if that's the case.
- [ ] Did you make sure to update the documentation with your changes? Here are the documentation guidelines, and here are tips on formatting docstrings.
- [ ] Did you write any new necessary tests?
Who can review?
Anyone in the community is free to review the PR once the tests have passed. Feel free to tag members/contributors who may be interested in your PR.
Overall looks pretty good !
I think we'll move to propre enum SYSTEM
instead of IS_XXXX
since there's no way a user could be running 2 devices simultaneously. But we'll take care of that in a subsequent PR if that's OK.
thanks very much @Narsil and @OlivierDehaene for your comment. adding @yao-matrix
@Narsil @OlivierDehaene I add xpu smi in env runtime, do you think is it proper to add this? this is to dump intel XPU version.
@Narsil @OlivierDehaene I add xpu smi in env runtime, do you think is it proper to add this? this is to dump intel XPU version.
It's OK in the current iteration (since showing N/A on most systems).
Feel free to open up the draft whenever you're OK so we can run the tests + merge.
@Narsil could you review the PR?done by my side.
@Narsil @mfuntowicz CI node is ready, could we move on?
@Narsil @mfuntowicz , pls feel free to let us know any issues, since CI is ready, I think we can launch CI and merge this PR, thx.
cc @kding1