llm icon indicating copy to clipboard operation
llm copied to clipboard

Access large language models from the command-line

Results 100 llm issues
Sort by recently updated
recently updated
newest added

This PR fixes the easier issues with #409, but does not completely resolve it. The `\n` vs `\r\n` is likely a red herring, as Python does a lot of newline...

> I'd like to fix those test failures, but I also want to ship a 0.13.1 release quickly. So I'm going to mark a bunch of tests as expected to...

help wanted
qa

This trick: https://til.simonwillison.net/pypi/pypi-releases-from-github

developer-experience

https://www.perplexity.ai/ https://api.perplexity.ai/chat/completions Perplexity is a popular AI service with its LLMs. Having it with llm as a plugin will add much value to the project. I tried to API into...

> I'm going with `llm collections show xxx` for consistency with `llm templates show xxx`. _Originally posted by @simonw in https://github.com/simonw/llm/issues/229#issuecomment-1712940173_

enhancement
embeddings

https://github.com/simonw/openai-to-sqlite/blob/361d98a7f260a1420e6e698481f298848b922253/README.md#saving-similarity-calculations-to-the-database This is the feature that can be used to save calculated similarity scores to the database. I use it to serve related TILs on my TILs site: https://til.simonwillison.net/llms/openai-embeddings-related-content ```bash...

enhancement
embeddings

Just noticed this: ```bash paginate-json 'https://api.github.com/repos/simonw/llm/issues?state=all&filter=all' > issues.json cat issues.json | jq '[.[] | {id, title, body}]' | llm embed-multi llm-issues - ``` That worked and populated my `llm-issues` collection...

bug
embeddings

Ollama makes it easy to run models such as llama2 locally on macOS easily: https://ollama.ai/ The user runs a server on localhost, so the architecture of the plugin could likely...

Hi, I'm seeing the following error when using the default model, 3.5: `Error: This model's maximum context length is 4097 tokens. However, your messages resulted in 8285 tokens. Please reduce...

Streaming response doesn't work with Azure models. This can be fixed by handling the JSON chunks differently. There is an issue with AzureChatOpenAI Streaming first response chunk response["choices"][0] is causing...