ragflow icon indicating copy to clipboard operation
ragflow copied to clipboard

[Question]: using minerU(vlm-vllm)

Open GXKIM opened this issue 1 week ago • 7 comments

Self Checks

  • [x] I have searched for existing issues search for existing issues, including closed ones.
  • [x] I confirm that I am using English to submit this report (Language Policy).
  • [x] Non-english title submitions will be closed directly ( 非英文标题的提交将会被直接关闭 ) (Language Policy).
  • [x] Please do not modify this template :) and fill in all the required fields.

Describe your problem

I set ‘vlm-transformers’ in my docker/.env file, but it’s still calling the default pipeline. Here are my logs. ###info Image Image ###postman Image

GXKIM avatar Dec 11 '25 08:12 GXKIM

To use MinerU with the vlm-transformers backend, you need to set the environment variable MINERU_BACKEND=vlm-transformers in your docker/.env file (not just 'vlm-transformers' by itself). After making this change, restart all Docker containers for the new setting to take effect. You can verify the active environment inside the container to ensure MINERU_BACKEND is set as expected. If it still defaults to the original pipeline, check for typos, duplicate/conflicting environment variable settings, or custom overrides in your deployment. The backend selection logic is based on MINERU_BACKEND, which supports 'pipeline', 'vlm-transformers', 'vlm-vllm-engine', and 'vlm-http-client' as valid options (docs/faq, mineru_parser.py).

To reply, just mention @dosu.


How did I do? Good | Irrelevant | Incorrect | Verbose | Hallucination | Report 🐛 | Other

Chat with Dosu Open in Cursor Join Discord Share on X

dosubot[bot] avatar Dec 11 '25 08:12 dosubot[bot]

Wait, that’s not right. vlm-transformers refers to an online model from Hugging Face. Since I’m running a model locally with LM Studio, should I be using vlm-http-client instead? This is assuming it can connect even though the service isn’t deployed via vllm.

env

Enable DocLing

USE_DOCLING=true

Enable Mineru

USE_MINERU=true MINERU_EXECUTABLE="$HOME/uv_tools/.venv/bin/mineru" MINERU_DELETE_OUTPUT=0 # keep output directory MINERU_BACKEND=vlm-http-client # or another backend you prefer MINERU_SERVER_URL=ip:11438

pptx support

DOTNET_SYSTEM_GLOBALIZATION_INVARIANT=1

Even after I made this change, it’s still using the YOLO model and not the multimodal model

GXKIM avatar Dec 11 '25 08:12 GXKIM

We are working on using mineru as one of the model providers like siliconflow. Soon will be released.

Magicbook1108 avatar Dec 11 '25 09:12 Magicbook1108

We are working on using mineru as one of the model providers like siliconflow. Soon will be released.

Will it support miner cluster configuration?

channlang avatar Dec 12 '25 01:12 channlang

Dose MinerU support cluster mode? Do you have reference about that?

KevinHuSh avatar Dec 12 '25 03:12 KevinHuSh

###env ‘USE_MINERU=true MINERU_EXECUTABLE="$HOME/uv_tools/.venv/bin/mineru" MINERU_DELETE_OUTPUT=0 # keep output directory MINERU_BACKEND=vlm-http-client # or another backend you prefer MINERU_SERVER_URL=http://10.xxxx:30000’

###miner u server

Image

I’m trying to use a VLM for multimodal parsing, but I’m unable to use the vlm-http-client. It seems I can’t connect to the service.

GXKIM avatar Dec 12 '25 08:12 GXKIM

Hi, @GXKIM

Which version are you using? I just tested it yesterday, it works well.

yongtenglei avatar Dec 12 '25 08:12 yongtenglei