OpenLLM
OpenLLM copied to clipboard
feat: Stream completions
Feature request
Is it possible to support stream completion similar to the OpenAI API?
Motivation
Can save users waiting time and improve user experience.
Other
No response
This has to do with SSE support on BentoML. There is a feature roadmap currently working on in the community discord.
This is now finished and supported