Eric Curtin
Eric Curtin
@ochafik kindly made the initial change to common_chat_templates_init in llama-run. This change ensures we don't pass nullptr to the function.
I merged this anyway to get rid of the seg fault people were reporting.
Incomplete and untested right now
I like the idea of using setfiles or something similar, less code to maintain... I think selinux=0, etc. is easy to handle in systemd .service file: ConditionKernelCommandLine=!selinux=0 We may be...
Hi @derrickqi2003 can you test if this change fixes your issue: https://github.com/containers/ramalama/pull/306 we haven't thoroughly tested docker support yet. Just podman support and without containers at all. It should only...
We added a basic version of whisper.cpp to the Container image here: https://github.com/containers/ramalama/pull/49
There's still more work for both `--runtime whisper.cpp` and `--runtime vllm` ... It can be tracked under this issue or somewhere else.
I would say just this to start: `ramalama --runtime whisper.cpp run ggml-large-v3-turbo.bin jfk.wav` which would just perform a fairly standard whisper.cpp command. No interactive support, it's not the same as...
@p5 kindly set up renovate for us.
I meant something like this, corrected: `ramalama --runtime whisper.cpp run ggml-large-v3-turbo.bin jfk.wav`