Pi

Results 108 comments of Pi

Llama.cpp context files are 1GB each, which is inconvenient to save in large amounts. It also takes longer to initialize the model.

Could you open a PR for this?

You need to download the q4_1 file, not q4_0.

> i tried so many models and they either fail to load or they never write anything at all, i used kobold and the models work fine so i dunno...

> from the link on the releases page > https://huggingface.co/Pi3141 And you're using q4_1, right?

> i tried this one > https://huggingface.co/Pi3141/gpt4-x-alpaca-native-13B-ggml/blob/main/ggml-model-q4_1.bin Can you try Alpaca native enhanced? https://huggingface.co/Pi3141/alpaca-7b-native-enhanced

> that one works, i guess its just really slow? also it doesnt seem to take into account other stuff that runs on my pc because, its running at 100%...

> i think you should add it or you are going to get more people reporting the models as broken Actually I can't. Llama.cpp doesn't show how many tokens of...

> bruh nobody is ever gonna wait one hour, they will just find another tool Yeah good luck to them finding a different tool thats faster than llama.cpp. If it...

> Sorry, I didn't test it on Mac before, I just assume when we run the command on terminal, it will display some info like this > ``` > >...