llama-cpp-python
llama-cpp-python copied to clipboard
Add support for auto setting n_gpu_layers from gguf and available vram size