OpenLLM
OpenLLM copied to clipboard
feat: Apple M1/M2 support through MPS
Feature request
I want to use OpenLLM with available models to run on Apple M1/M2 processors (GPU support) through MPS.
Today:
openllm start falcon
No GPU available, therefore this command is disabled
Motivation
No response
Other
No response
I'm currently disabling falcon on MPS since I would just run out of memory to try even run the model on Mac
Not sure if this is valid any more. I have since tested a lot with pytorch on MPS, and it is often slower. Will probably investigate mlc vs. gguf for this.