M Waleed Kadous
M Waleed Kadous
Great suggestion! We looked at this. This turned out to be too limiting (e.g. Aviary has support for optimized batching, whereas OpenAI's GPT-3.5-Turbo interface does not). But we are planning...
PR for option 1 at https://github.com/hwchase17/langchain/pull/5661 (@hwchase17 jfyi). Option 3 merged at https://github.com/ray-project/aviary/commit/8e4e965bb19e7944f9687d9b89b4e47d4aa069d0
Could we please check this in to a repo as a supported LLM config?
@Yard1 what do you think about making the frontend run on port 7860 by default to be consistent with normal Gradio and not cause this problem?
Will prioritize fixing this ASAP. Thank you for reporting the bug!
Hey Theo, Thanks for reaching out! The point is you don't use an LM, you use a vector database. You embed questions like capitals in a semantic search index. This...
Weirdly, I tried it myself and it was considerably slower: like 20x slower. But I think that would be a really good section to add, especially with us also adding...