ragflow icon indicating copy to clipboard operation
ragflow copied to clipboard

[Question]: Inconsistent AI Response Lengths between Ragflow and Direct ollama Interactions

Open lzivan opened this issue 1 year ago • 11 comments

Describe your problem

Hello Ragflow team,

I am encountering an issue where the AI responses I receive when using ollama through Ragflow are significantly shorter and less detailed compared to when I interact directly with ollama.

Could you please help understand why there is such a discrepancy and how it might be resolved?

image image

lzivan avatar Jun 13 '24 09:06 lzivan

I'm encountering the same issue. Docker dev installed yesterday. Responses are cut off. I increased output tokens to 2048 but that doesn't make any difference. image

Tried with several ollama models (llama3, mistral-openorca, llama3-chatqa) with same truncated results.

Furthermore and related to this, RAGflow says that the Knowledge Base does not provide information, but it identifies the correct file and even the correct section as source for the response.

AnonTester avatar Jun 13 '24 15:06 AnonTester

I'm encountering the same issue. Docker dev installed yesterday. Responses are cut off. I increased output tokens to 2048 but that doesn't make any difference. image

Tried with several ollama models (llama3, mistral-openorca, llama3-chatqa) with same truncated results.

Furthermore and related to this, RAGflow says that the Knowledge Base does not provide information, but it identifies the correct file and even the correct section as source for the response.

If it refers to correct chunks (you can test it in retrieval test) but the answer is still about nothing, it's probably caused by the LLM since its capability shortage to under stand the relevance between question and chunks.

KevinHuSh avatar Jun 14 '24 01:06 KevinHuSh

Describe your problem

Hello Ragflow team,

I am encountering an issue where the AI responses I receive when using ollama through Ragflow are significantly shorter and less detailed compared to when I interact directly with ollama.

Could you please help understand why there is such a discrepancy and how it might be resolved?

image image

What about disable max token toggle in dialog setting?

KevinHuSh avatar Jun 14 '24 01:06 KevinHuSh

I disabled the "max token toggle" as mentioned and it worked. Thanks!

lzivan avatar Jun 14 '24 02:06 lzivan

Worked for me as well and got a full and proper response from the LLM to the same question now. Thanks!

AnonTester avatar Jun 14 '24 03:06 AnonTester

max token toggle

the images can't see now, where I can set the 'max token toggle'?

colin4k avatar Jun 17 '24 02:06 colin4k

max token toggle

the images can't see now, where I can set the 'max token toggle'?

Chat properties -> model settings -> toggle switch at bottom: Screenshot_2024-06-17-18-39-34-15_40deb401b9ffe8e1df2f1cc5ba480b12

AnonTester avatar Jun 17 '24 17:06 AnonTester

Describe your problem

Hello Ragflow team,

I am encountering an issue where the AI responses I receive when using ollama through Ragflow are significantly shorter and less detailed compared to when I interact directly with ollama.

Could you please help understand why there is such a discrepancy and how it might be resolved?

image image

btw, did you running ollama in Macosx ?

yangboz avatar Jun 18 '24 05:06 yangboz

max token toggle

the images can't see now, where I can set the 'max token toggle'?

Chat properties -> model settings -> toggle switch at bottom: Screenshot_2024-06-17-18-39-34-15_40deb401b9ffe8e1df2f1cc5ba480b12

got it, thank u.

colin4k avatar Jun 20 '24 05:06 colin4k

max token toggle

the images can't see now, where I can set the 'max token toggle'?

Chat properties -> model settings -> toggle switch at bottom: Screenshot_2024-06-17-18-39-34-15_40deb401b9ffe8e1df2f1cc5ba480b12

does work!!

tyro12138 avatar Jun 27 '24 05:06 tyro12138

Yes, "Chat properties -> model settings -> toggle switch at bottom" helps! Thanks!

netandreus avatar Jul 09 '24 07:07 netandreus

Yes, "Chat properties -> model settings -> toggle switch at bottom" helps! Thanks!

How about in agent? Same issue found in agent.

sliontc avatar Jan 07 '25 03:01 sliontc

same issue in agent. In chat response is big, but agent response is limited.

martinarisk avatar Apr 14 '25 08:04 martinarisk

when testing 0.17.2 slim Chat's response is big, but agent embedded to website response is limited (truncated).

martinarisk avatar Apr 14 '25 08:04 martinarisk

Note the LLM is claude 3.5 and provider is AWS Bedrock

martinarisk avatar Apr 14 '25 10:04 martinarisk

I was reading pull request 845. There Max Tokens is avaliable for deepseek-chat. But I don't have this option in bedrock:

Image

martinarisk avatar Apr 14 '25 10:04 martinarisk