Saifeddine ALOUI
Saifeddine ALOUI
Thank you for answering. I have submitted an issue on this before starting to code: [#1135](https://github.com/LAION-AI/Open-Assistant/issues/1135)
I have already done this to another project. I have done a github action that automatically translates a csv file containing the data I was adding into other languages automatically....
But I still think the idea of translating prompts and answers to other languages may be interesting. I had this idea when I was adding french prompts. I said to...
OK we can close it. I have another pending pull request for the automated translation part [here](https://github.com/LAION-AI/Open-Assistant/pull/1394#)
Thank you very much. I didn't seer that :)
If you have enough VRAM to run multiple models, you can create multiple instances of ollama with different port numbers then use my proxy to manage access and route to...
That's why I had to build a proxy. You can install multiple servers on a single or multiple machines then use my proxy to service multiple users with a multi...
well technically, you can run multiple instances of the same model by runnin multiple instances of ollama with different port numbers, configure them in the proxy config file then they...
ollama does not support batching and unless they build something like vllm, I don't see how you can do it.
> The idea is the system (a MacBook or Mac studio with 128gb+ of unified RAM) can handle several dozen 4-bit 7b models at once. > > So taking more...