ragflow icon indicating copy to clipboard operation
ragflow copied to clipboard

[Question]: Difference between Local LLM and Online LLM

Open TomTigerDong opened this issue 11 months ago • 3 comments

Describe your problem

Ragflow calls the large model deployed locally with Ollama, and it responds to questions very quickly. However, when calling the large model API, it seems to be lost somewhere, resulting in a very slow response. Moreover, it appears to "dumb down" the large model, and the answers are quite silly. ——— This issue has been resolved. When calling the external large model API, Ragflow hides the thinking process, which is why it seems slow. The results are actually consistent with what is expected, and there is no "dumbing down" of the model.

TomTigerDong avatar Feb 19 '25 09:02 TomTigerDong

We intend to create an international community, so we encourage using English for communication.

JinHai-CN avatar Feb 19 '25 13:02 JinHai-CN

Different large models have different capabilities, and local deployed large models and online large models also vary in terms of the accuracy of answering questions and latency.

JinHai-CN avatar Feb 19 '25 13:02 JinHai-CN

——— This issue has been resolved. When calling the external large model API, Ragflow hides the thinking process, which is why it seems slow. The results are actually consistent with what is expected, and there is no "dumbing down" of the model.

TomTigerDong avatar Mar 04 '25 05:03 TomTigerDong