phalexo

Results 137 comments of phalexo

This makes the most sense. Llama.cpp is able to load at least one of the models I tried. Is there an easy way to pull that specific commit? On Thu,...

Temporary work-around is to set the context to 60000. Not as good as 128K, but better than 4K. This appears to work for ollama with my four(4) 12.2GiB Titan GPUs....

> Similar issue here with ollama v0.1.27 and mistral. Upgraded to ollama v0.1.29 same problem. Pulled codellama and reconfigured pythagora to use this model....same problem I also tried with LM...

The LLMs often respond with some comments, in addition to JSON formatted text, and then these comments trigger this error. I think there has to be a step filtering content...

@ht-it > I had the same JSON issue with Mistral 7b, but I managed to resolve it using codellama 13b. > > However, prompt processing issues sometimes cause request errors....

If someone wants to play with prompts and ollama. This is what I have for deepseek 33B model. Maybe you can improve it and make it more reliable. FROM /opt/data/TheBloke/deepseek-coder-33B-instruct-GGUF/deepseek-coder-33b-instruct.Q6_K.gguf...

New and different errors. This is with Mixtral 8x7B-instruct { "steps": [ { "type": "command", "command": { "command": "npm install", "timeout": 60000, "success\_message": "successfully installed" } }, { "type": "command",...

@tescolopio Can you clarify what you mean by "working?" It is pretty easy to get the CLI version to work with ollama OpenAI's API. The problem for me arises when...

I wonder if someone could put together a web scraper API to talk to chatGPT/GPT-4 via a browser? Then it would be $20/month. It would be throttled of course and...

After a bit of thinking I've come to the following conclusions: 1) prompts are not universal, they match specific models. So, each model needs its own prompts. 2) One can...