Steve
Steve
for the minimum size, it seems the k/m is not supported; maybe we should get an error message if a letter is used or it should be supported?
This is what I did, so not sure which step I missed? I understood also that we refer to the encoder, not GenerativeAI (and experience could be lower), but I...
For scalability reason, running the embeddings within the main platform is a problem; by itself, danswer doesn't need GPU and running embeddings is a low priority vs responsiveness. Maybe using...
for the point 1, it doesn't seem obvious because storm is calling the endpoint /v1/completions and not /v1/chat/completions so direct ollama is out-of-question. Maybe litellm could help.
yes I can help. Today, I am using litellm (through docker) to ollama or the ollama openAI endpoint directly. What is important is to be able to change the base_url...
Ollama provides an openAI compatible endpoint. It could make sense to focus on this one or you have a concern? I was running some tests with mistral and wizardlm2; llama3...
to run: `python examples/run_storm_wiki_mistral.py --url http://192.168.0.120 --port 4000 --do-generate-outline --remove-duplicate --do-research`, I had to add in the requirements.txt: anthropic and streamlit Also ideally the model name is an argument as...
Following my experience above: `python examples/run_storm_wiki_mistral.py --url http://192.168.0.120 --port 4000 --do-generate-outline --remove-duplicate --do-research` I am getting the error message below and I am not sure how to debug it further:...
> @shuther The error `root : ERROR : Error occurs when searching query : 'hits'` you see is most likely due to rate limit exceeds from You.com search engine API....
I did git pull then I have this: http://192.168.0.120 is an openAI compatible endpoint (litellm) python examples/run_storm_wiki_mistral.py --url http://192.168.0.120 --port 4000 --max-thread-num 1 --do-generate-outline --do-research Traceback (most recent call last):...