Jaeger M
Jaeger M
Tried on Chronos 13b, wizard-lm 13b, and Pygmalion 7b, I used temperatures in between 0.5 and 1, and a context length of 2048, lower temperatures do seem to wrangle it...
occam's fork of koboldAI that allows using exllama using gptQ, said fork behaves normally
you think maybe the code wasn't hooked up to the context correctly and it's actually running on incredibly low context size?
Alright then, thank you for taking a look at it
I can confiirm my issue is no longer present after Occam's latest commit to his koboldAI fork, thank you very much for your help.
so does this mean the fix has been rolled into the code, and if so, what files do I replace?
I will try that when I get a chance to, thank you