fastLLaMa icon indicating copy to clipboard operation
fastLLaMa copied to clipboard

GGUF and/or LLama-3 support?

Open BHX2 opened this issue 1 year ago • 0 comments

Is this project updated enough to use gguf files or the LLama-3 architecture? I see that the documentation examples use ggml via .bin files which I'm assuming was the previous file format. I'm specifically interested in the loading / unloading feature for LoRa feature that doesn't seem supported in llama.cpp by itself

BHX2 avatar Apr 30 '24 00:04 BHX2