ragflow icon indicating copy to clipboard operation
ragflow copied to clipboard

[Feature Request]: LMstudio rerank model isn't supported

Open r3vs opened this issue 7 months ago • 2 comments

Self Checks

  • [x] I have searched for existing issues search for existing issues, including closed ones.
  • [x] I confirm that I am using English to submit this report (Language Policy).
  • [x] Non-english title submitions will be closed directly ( 非英文标题的提交将会被直接关闭 ) (Language Policy).
  • [x] Please do not modify this template :) and fill in all the required fields.

RAGFlow workspace code commit ID

RAGFlow image version

v0.19.0-full

Other environment information

Windows 11 with Docker

Actual behavior

Image

This is the log:

2025-06-07 00:45:35,419 ERROR 22

Fail to access model(text-embedding-qwen.qwen3-reranker-0.6b).The LmStudioRerank has not been implement

NoneType: None

2025-06-07 00:45:35,421 INFO 22 172.20.0.6 - - [07/Jun/2025 00:45:35] "POST /v1/llm/add_llm HTTP/1.1" 200 -

2025-06-07 00:45:37,097 ERROR 22

Fail to access model(text-embedding-qwen.qwen3-reranker-0.6b).The LmStudioRerank has not been implement

NoneType: None

2025-06-07 00:45:37,098 INFO 22 172.20.0.6 - - [07/Jun/2025 00:45:37] "POST /v1/llm/add_llm HTTP/1.1" 200 -

Expected behavior

The docs says that reranking models for LMstudio are supported.

Steps to reproduce

Image

Additional information

No response

r3vs avatar Jun 06 '25 16:06 r3vs

Hi @r3vs! I'm Dosu and I’m helping the ragflow team.

The LMstudio rerank model is listed in the codebase, but its implementation is missing—the main method raises a "NotImplementedError" and isn't functional yet, which matches the error you're seeing in your logs. This means that, despite being recognized as a model type, LMstudio rerank is not actually supported in RAGFlow at this time see source.

There are also similar reports for other rerank models not being supported yet. As a workaround, some users have used LM_Studio to add a local reranker model or selected one from the FlagEmbedding repository see discussion.

I couldn't find documentation that explicitly states LMstudio rerank models are supported, so there may be a mismatch between the docs and the current implementation. If you have a specific doc reference, sharing it could help clarify the discrepancy.

To reply, just mention @dosu.


How did I do? Good | Irrelevant | Incorrect | Verbose | Hallucination | Report 🐛 | Other  Join Discord Share on X

dosubot[bot] avatar Jun 06 '25 16:06 dosubot[bot]

@r3vs Noted

ZhenhangTung avatar Jun 11 '25 03:06 ZhenhangTung

Any update on this? I got this error too, but the supported model list is contradicting this issue: https://ragflow.io/docs/dev/supported_models

joaorsbarbosa avatar Sep 07 '25 16:09 joaorsbarbosa

Any update on this? I got this error too, but the supported model list is contradicting this issue: https://ragflow.io/docs/dev/supported_models

@joaorsbarbosa which version are you using?

ZhenhangTung avatar Sep 09 '25 03:09 ZhenhangTung

@joaorsbarbosa which version are you using?

v0.20.4 Although we have talked on Discord already! I can quote your reply here if you want, for reference sake.

joaorsbarbosa avatar Sep 09 '25 08:09 joaorsbarbosa

@joaorsbarbosa which version are you using?

v0.20.4 Although we have talked on Discord already! I can quote your reply here if you want, for reference sake.

@joaorsbarbosa Sure, if it’s not too much trouble for you 🙂.

ZhenhangTung avatar Sep 09 '25 11:09 ZhenhangTung

For reference:

Haven't planned it yet. We're refactoring our pile parsing pipeline right now. A feature request is welcomed.

PS: Will submit that feature request in a bit! 😅

joaorsbarbosa avatar Sep 09 '25 12:09 joaorsbarbosa

The LM Studio API endpoint follows the OpenAI-compatible format (/v1/chat/completions, /v1/completions, /v1/embeddings), which supports chat, vision, and embedding models.

However, LM Studio does not implement a dedicated /v1/rerank endpoint or reranking logic. Even if a reranker model is loaded, it is still handled as a chat model through the standard LLM pipeline. Using a rerank model via the chat interface is therefore conceptually misaligned—it asks a ranking model to generate text instead of computing relevance scores, leading to unstable or meaningless results.

Magicbook1108 avatar Oct 13 '25 03:10 Magicbook1108