Vincent Min
Vincent Min
I don't know the specifics of how Ollama achieves JSON mode, but let me point out that vLLM supports [outlines](https://github.com/outlines-dev/outlines) and [lm-format-enforcer](https://github.com/noamgat/lm-format-enforcer) for guiding generation, [see the vLLM docs](https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#extra-parameters-for-chat-api). It...
I can also confirm that Ollama embeddings for `snowflake-arctic-embed:137m-m-long-fp16` are not behaving as expected. I set up a synthetic benchmark for internal testing. I take 500 articles and use an...
We are suffering from the same issue. Indeed the problem occurs stochastically, preventing us from creating a minimal example. We are using scoop together with the DEAP package.
An official lm-evaluation-harness Docker image would be great. I just found that BigScience hosts the following Docker image: https://github.com/orgs/bigcode-project/packages/container/package/evaluation-harness
For anyone interested, here's a clean solution: ```python from langsmith import tracing_context llm = ChatOpenAI() with tracing_context(enabled=False): # Anything in this code block will **not** be traced to LangSmith llm.invoke("hello...
I encountered this issue too after migration to `langchain-core==0.3.x` and switching from `Pydantic` v1 to v2. It would be great if Ragas could be updated to be compatible.
Thanks for clarifying, Nathan. I was assuming that the LLM was used only to decide which agent should be used to address the current question, i.e. the LLM functions purely...
Yes, that's correct.
This feature is a must have. I am interested in using Llamafile as an alternative to Ollama, but lack of support for prompt templates is a dealbreaker.