ai2p
ai2p
indeed, how?
> "### Human:" How to delete "### Human:" part from Vicuna's response?
I've been able to compile latest standard llama.cpp with cmake under the Windows 10, then run[ ggml-vicuna-7b-4bit-rev1.bin](https://huggingface.co/eachadea/ggml-vicuna-7b-4bit/tree/main) , and even [ggml-vicuna-13b-4bit-rev1.bin](https://huggingface.co/eachadea/ggml-vicuna-13b-4bit/tree/main), with this command-line code (assuming that your .bin in...
> Context swap. The context fills up and then the first half of it is deleted to make more room, but that means that the whole context has to be...
> can someone explain to me what the difference between these two options is? Same except temperature and amount of threads. About temperature [read here](https://medium.com/@lazyprogrammerofficial/what-is-temperature-in-nlp-llms-aa2a7212e687). As I preliminary think, the...
> Therefore I made the following quick fix for vicuna: I've repeated your modifications, but nothing changes - it still shows "### Human:" each time... Who knows how to make...
> The vicuna v1.1 model used a different setup. Uhhh... Such a mess... Definitely needed some standardization for peple teaching LLMs! At least, with tokens such assistant/human/eos it should be...