thepipe icon indicating copy to clipboard operation
thepipe copied to clipboard

Colaboration for Ollama / Gemini support

Open patrickab opened this issue 2 weeks ago • 0 comments

Hi,

I am currently working on personal toolkit for LLMs from diverse providers & while searching for effective methods to mine large .pdf documents (eg books, pdf slides) Ive come across your project.

I am particularly interested in extending the functionality of scraping PDFs with other VLM Clients (especially locally hosted Ollama models). Instead of forking this project & building this for myself I thought it'd be nice to integrate this functionality into this project for others.

I've alread looked into your project & it would be necessary to generalize

def scrape_pdf(
    file_path: str,
    openai_client: Optional[OpenAI] = None,

to a VLM Wrapper Class, that handles the different syntaxes for Ollama / OpenAI / Gemini

def scrape_pdf(
    file_path: str,
    vlm_client: Optional[VLMClient] = None,

I already have a finished LLMClient wrapper that wraps multimodal capabilites for all mentioned Clients into a single object. If you are interested in integrating this functionality I would be happy to cooperate.

patrickab avatar Nov 22 '25 17:11 patrickab