Antonino Lorenzo

Results 10 comments of Antonino Lorenzo

@cyborgmarina I was thinking exactly about not using `litellm`. I am working on the `ollama-support` branch on my own fork given that I am not one of the repository main...

@Mihir1003 @killind-dev Thanks for the support, I made the required changes to `OllamaModel` class and tested the system prompt within `ollama_prompts.py`; it seems to work correctly, however to ensure correct...

Hi, when you provide tools to `chat` and set `stream=True` you'll either get a normal streams if the model doesn't call any tool, or, in your case, you should get...

I also encountered this issue, however rather than stopping I think a throttling mechanism would be more useful. If you are going to fix this you should consider: - throttling...

Man before you opened that issue I didn't knew I could use tools 💀 Maybe I am the one skipping docs, however I would like to know where/how did you...

Ok, however I wrote some code to see how it could be done, there are still some edge cases to handle and I wrote it kind of in a rush,...

> In [open webui tools.py file](https://github.com/open-webui/open-webui/blob/main/backend/utils/tools.py#L40) they just write "object": I misinterpreted its meaning. As shown it is simple to implement, I would enjoy working on it myself, however there...

> In my opinion download a whole framework for something so basic like that (langchain) is not interesting. I agree, I rarely use langchain and I do not feel comfortable...

Technically there is no way of doing it, as you can see in that [issue](https://github.com/ggerganov/llama.cpp/discussions/6212) on `llama.cpp` repository; the "brute-force" way is to add a `KeyboardInterrupt` during the stream, for...

@Jaswanth-Pillella how are you running it Google Colab?