llama-gpt icon indicating copy to clipboard operation
llama-gpt copied to clipboard

[Feature Request] Add AMD ROCm Support

Open cotsuka opened this issue 2 years ago • 7 comments

llama.cpp supports ROCm, which would open this project to be used on AMD hardware as well: https://github.com/ggerganov/llama.cpp#hipblas

cotsuka avatar Aug 29 '23 01:08 cotsuka

Happy to test on a RX 6800 XT if needed.

agates avatar Aug 30 '23 15:08 agates

Also willing to test, have a 6600 and a 6400 here

hydragyrum32 avatar Sep 01 '23 23:09 hydragyrum32

+1 for testing this, I have a 6900XT

tagyro avatar Sep 09 '23 10:09 tagyro

Willing to help testing with my 7900XTX

AnttiRae avatar Sep 27 '23 12:09 AnttiRae

Possibly quite simple to implement as there are good examples for CUDA https://github.com/getumbrel/llama-gpt/tree/master/cuda

jykae avatar Sep 27 '23 13:09 jykae

Hey @jykae @AnttiRae @tagyro @hydragyrum32 @agates @cotsuka .

https://github.com/getumbrel/llama-gpt/pull/114

I need help testing this, I don't have an AMD GPU but I've created a POC for AMD Support. Please test if possible and let me know in the comments so I can make the required changes.

ParthJadhav avatar Oct 18 '23 04:10 ParthJadhav

@AnttiRae check out, looks like there's candidate for ROCm support now 😊

jykae avatar Oct 23 '23 15:10 jykae