llm_aided_ocr icon indicating copy to clipboard operation
llm_aided_ocr copied to clipboard

Enhance Tesseract OCR output for scanned PDFs by applying Large Language Model (LLM) corrections.

Results 15 llm_aided_ocr issues
Sort by recently updated
recently updated
newest added

Hi, Your code used llma2 chat offline LLM. But, I wanted to use alternative offline LLMs such as huggingface's distilbert or roberta or albert. Do you have any suggestion for...

Is there any plan to restructure the code to be uniform to use it with Llama2/API like (gpt-3.5-turbo, gpt-4) to use this PDF-to-text in any hardware. https://github.com/Dicklesworthstone/llama2_aided_tesseract/blob/5719a9aede6b0666f6f08d239cac7b1550298b79/tesseract_with_llama2_corrections.py#L180 https://github.com/Dicklesworthstone/llama2_aided_tesseract/blob/5719a9aede6b0666f6f08d239cac7b1550298b79/tesseract_with_llama2_corrections.py#L122 https://github.com/Dicklesworthstone/llama2_aided_tesseract/blob/5719a9aede6b0666f6f08d239cac7b1550298b79/tesseract_with_llama2_corrections.py#L173

for these `doc format convertion`, `text summarization` tasks, I think one of key feature is to include all or some of the images/charts/tables from original doc, as those elements often...

Your provided [tesseract_with_llama2_corrections.py] code snippet is equipped with the llma2 chat ggml q3 k_s.bin LLM model but the huggingface.co is referring to use GGUF instead saying the GGML is deprecated....

ERROR: ERROR: Failed to build installable wheels for some pyproject.toml based projects (llama-cpp-python) I have updated the gcc version but still the error. Working for ubuntu verison 24.04 LTS but...

I am using a 16 core cpu for the same document using a local model same as in the github repo. How to get the output faster?

Token indices sequence length is longer than the specified maximum sequence length for this model (2816 > 2048). Running this sequence through the model will result in indexing errors 2024-08-13...

I am getting negative max tokens in AI call.. Some small bug probably... ``` 2024-08-12 09:54:03,183 - ERROR - An error occurred while processing a chunk: Error code: 400 -...

It would be super nice to integrate this with https://github.com/ocrmypdf/OCRmyPDF

Even after properly following the instructions, it throws error: ValueError: Failed to load model from file: .Llama-2-13B-chat-GGML/llama-2-13b-chat.ggmlv3.q4_K_S.bin I have checked with all other model files, same error persists