docker-faster-whisper icon indicating copy to clipboard operation
docker-faster-whisper copied to clipboard

[FEAT] AMD GPU ROCm Support

Open Teejer opened this issue 6 months ago • 10 comments

Is this a new feature request?

  • [x] I have searched the existing issues

Wanted change

I would like another image with AMD ROCm support added. To make it faster on amd machines.

Thank you!

Reason for change

to make whisper faster on amd machines.

Proposed code change

No response

Teejer avatar Jun 11 '25 18:06 Teejer

Thanks for opening your first issue here! Be sure to follow the relevant issue templates, or risk having this issue marked as invalid.

github-actions[bot] avatar Jun 11 '25 18:06 github-actions[bot]

This issue has been automatically marked as stale because it has not had recent activity. This might be due to missing feedback from OP. It will be closed if no further activity occurs. Thank you for your contributions.

LinuxServer-CI avatar Jul 12 '25 10:07 LinuxServer-CI

Does this sound like something you guys plan on doing at some time?

Teejer avatar Jul 14 '25 14:07 Teejer

rocm is like cuda it is massive and it needs to be inside the container there is no runtime to pass the host install to the container, its like 4 gigs most people just base on their baseimages like rocm/rocm-terminal.

So basically no, we are not planning on adding 4 gigs to either the baseimages or this container, it sucks that this is how it is implemented and it walls stuff out, but maybe a docker mod that runs on init ? Just spitballing a possible solution.

thelamer avatar Jul 14 '25 14:07 thelamer

Could you not make a :rocm container one, and base it on your base? So then its not making everyone's images that much larger? Then us with amd can get the rocm image.

Teejer avatar Jul 14 '25 15:07 Teejer

Sure it is possible but I am just chiming in as the why as a team member. I do not maintain this image and I do not use this program, just thought I would give you information.

You also realize that doing this without a cuda one is pointeless as nvidia has something like 90% of the GPU share?

You starting to see how this becomes a big cumbersome thing to support ?

thelamer avatar Jul 14 '25 15:07 thelamer

This issue has been automatically marked as stale because it has not had recent activity. This might be due to missing feedback from OP. It will be closed if no further activity occurs. Thank you for your contributions.

LinuxServer-CI avatar Aug 15 '25 10:08 LinuxServer-CI

I'll consider it, but I need to do a bunch of testing because I have no idea what's actually required to support rocm with faster-whisper and I've only got a RX 7700S to play with, which doesn't technically have rocm support (even though it seems to mostly work).

thespad avatar Aug 20 '25 15:08 thespad

localai added gpu support for faster-whisper:

Re.: https://github.com/mudler/LocalAI?tab=readme-ov-file#hardware-acceleration-matrix

They have a model ready to download and the base also ready.

Repo with igpu intel: https://github.com/tannisroot/wyoming-whisper-cpp-intel-gpu-docker

I also find this: https://github.com/roryeckel/wyoming_openai

I will test and switch for one that I was able to use openai API and wyoming to use only just one endpoint. Maybe consider one of these to switch in future.

webysther avatar Nov 11 '25 13:11 webysther

LocalAI after don't worked as expected, lots of missings parts, but promissing... wyoming-whisper-cpp-intel-gpu-docker is abandoned... wyoming_openai works but no good candidate that really uses well the llama.cpp...

webysther avatar Nov 11 '25 19:11 webysther