ai123
ai123
Thanks. pls review the information as followed (add OLLAMA_DEBUG=1 to my server environment)  ### FastGPT log (to make it easier for you to understand, I have translated some of...
Thanks for your replay. ollama log file attached,pls. Fell free to let me know if more information needed... [ollama.log](https://github.com/user-attachments/files/17352852/ollama.log) You can start by reviewing the following conten: 10月 13 10:50:10...
Thanks for your information, I almost understand the root cause--- the token qty. which LLM can supported. The following was testing with glm4:9b, which context length 131072. Whatever setting --ctx-size...
Thank you for your response. I understand your explanation. According to your explanation, both models eventually ran successfully, but why didn't either model output the results as instructed? Could you...
I have found the root cause of the problem. The reason for the unsuccessful output was that the connect_time and other parameters were not set in nginx. The issue is...
> You have `OLLAMA_NUM_PARALLEL` unset. In ollama-5.log, you set a context size of 96001. Since you have lots of VRAM, ollama first tries to load the model with `OLLAMA_NUM_PARALLEL=4`. That...
Thank you for your reply. I have conducted the following two tests, and I am unsure how to handle some issues, so I need your further assistance. ## 1. Overall...
一直开着 ---原始邮件--- 发件人: ***@***.***> 发送时间: 2025年4月14日(周一) 下午2:36 收件人: ***@***.***>; 抄送: ***@***.******@***.***>; 主题: Re: [labring/FastGPT] qwq在批量执行控件中报错 (Issue #4484) 流式输出的开关打开试试看 — Reply to this email directly, view it on GitHub, or unsubscribe....
一直都用的流 ---原始邮件--- 发件人: ***@***.***> 发送时间: 2025年4月17日(周四) 晚上11:27 收件人: ***@***.***>; 抄送: ***@***.******@***.***>; 主题: Re: [labring/FastGPT] qwq在批量执行控件中报错 (Issue #4484) 这个模型只支持流,不支持非流 — Reply to this email directly, view it on GitHub, or unsubscribe....
+1