crawl4ai icon indicating copy to clipboard operation
crawl4ai copied to clipboard

[Bug]: Docker API LLM Operations / Extractions

Open Codinator007 opened this issue 3 weeks ago • 1 comments

crawl4ai version

0.7.6

Expected Behavior

Same as other people I am not able to get llm extraction to run. I have setup via docker compose and want to use for llm extraction my local (same vm) vllm or ollama endpoints. I wish I could turn on LiteLLM Debug Logs because right now i only see this in the logs:

stage_crawl4ai-1  | LiteLLM.Info: If you need to debug this error, use `litellm._turn_on_debug()'.
stage_crawl4ai-1  | 
stage_crawl4ai-1  | 
stage_crawl4ai-1  | Provider List: https://docs.litellm.ai/docs/providers

I tried nearly everything:

  • config.yml with llm config
  • .llm.env with custom LLM_BASE_URL
  • as Provider i tried
    • LLM_PROVIDER=ollama/gpt-oss-20b
    • LLM_PROVIDER=hosted_vllm/gpt-oss-20b
    • LLM_PROVIDER=hosted_vllm/openai/gpt-oss-20b
    • etc.

Current Behavior

no debug logs of litellm extraction with llm not work. for example i use /md endpoint with f:llm & q:Extract only x

Is this reproducible?

Yes

Inputs Causing the Bug


Steps to Reproduce


Code snippets


OS

docker ubuntu 22.04

Python version

3.10.12 on the host

Browser

No response

Browser version

No response

Error logs & Screenshots (if applicable)

No response

Codinator007 avatar Nov 13 '25 14:11 Codinator007