Ernest O'Connell

Results 2 comments of Ernest O'Connell

Thx for your reply, but the `8k > max_token(450) + prefix_prompt(about 100) + content(about 2000)`。 Unfortunately,I reduce max_tokens to 50, it still doesn’t work。 I have no problem interacting with...

> In my scene,the backend model is xinrerence and chatglm3-6b-**32k** . dify max_tokens sets to 500 It's no problems when interacting with xinference , content length 2164 Chinese characters (include...