Todsaporn Banjerdkit
Todsaporn Banjerdkit
> Can somebody list down the steps to override the tokenizer file. I am new to this. Need Help! Form my understand is 1. download tokenizer file from original repo...
> Can you perhaps try this? > > https://docs.gaianet.ai/creator-guide/knowledge/text This one took 6.88s seem to be faster.🤔
I'm not quite sure which line i've to check, I follow step from readme which is ```curl wasmedge --dir .:. --nn-preload default:GGML:AUTO:Llama-2-7b-chat-hf-Q5_K_M.gguf \ --nn-preload embedding:GGML:AUTO:all-MiniLM-L6-v2-ggml-model-f16.gguf \ rag-api-server.wasm \ --model-name Llama-2-7b-chat-hf-Q5_K_M,all-MiniLM-L6-v2-ggml-model-f16...
Yes i just run 100% steps in the docs(for many times by now), but it's still slow. I think I miss something pretty obvious 🤔.
Hey, Thanks for response, Here's full log (still running) ``` source /home/katopz/anypost/.venv/bin/activate (base) katopz@shikuwa:~/anypost$ source /home/katopz/anypost/.venv/bin/activate (anypost) (base) katopz@shikuwa:~/anypost$ uv run 2048.py INFO 09-13 13:07:15 [__init__.py:235] Automatically detected platform cuda....
Will do, `wonnx` and `ort` is on my waiting list. Anyway yesterday I try explore/build/compile native/wasm examples from https://github.com/huggingface/candle (yes i still evaluate things here) I like to know what...
Sorry, to say but not real soon because 1. I still has no idea how to accomplish that yet. 2. I will get fire in the next 2 months so...
 Qwen 2.5 Coder 7B outperforms many model there, so yes this is a must have.
a bit confuse here, do i need to wait for the PR to merge or use openai workaround?
Thanks, that's understandable. Anyhow this (working code) took me sometime to figure this out 😅 so maybe this should be add to the docs somewhere. ```rust let client = Client::builder("")...